2025-01-08 04:57:58,780 main DEBUG Apache Log4j Core 2.17.2 initializing configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@4b8d604b 2025-01-08 04:57:58,811 main DEBUG Took 0.027981 seconds to load 1 plugins from package org.apache.hadoop.hbase.logging 2025-01-08 04:57:58,812 main DEBUG PluginManager 'Core' found 129 plugins 2025-01-08 04:57:58,812 main DEBUG PluginManager 'Level' found 0 plugins 2025-01-08 04:57:58,814 main DEBUG PluginManager 'Lookup' found 16 plugins 2025-01-08 04:57:58,815 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 04:57:58,824 main DEBUG PluginManager 'TypeConverter' found 26 plugins 2025-01-08 04:57:58,844 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.util.MBeans", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 04:57:58,847 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 04:57:58,848 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.logging.TestJul2Slf4j", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 04:57:58,848 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 04:57:58,849 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.zookeeper", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 04:57:58,849 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 04:57:58,850 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSinkAdapter", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 04:57:58,851 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 04:57:58,851 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSystemImpl", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 04:57:58,852 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 04:57:58,853 main DEBUG LoggerConfig$Builder(additivity="false", level="WARN", levelAndRefs="null", name="org.apache.directory", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 04:57:58,853 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 04:57:58,854 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.ipc.FailedServers", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 04:57:58,857 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 04:57:58,858 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsConfig", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 04:57:58,859 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 04:57:58,859 main DEBUG LoggerConfig$Builder(additivity="null", level="INFO", levelAndRefs="null", name="org.apache.hadoop.hbase.ScheduledChore", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 04:57:58,860 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 04:57:58,861 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.regionserver.RSRpcServices", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 04:57:58,861 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 04:57:58,862 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 04:57:58,865 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 04:57:58,866 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 04:57:58,866 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 04:57:58,867 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hbase.thirdparty.io.netty.channel", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 04:57:58,867 main DEBUG Building Plugin[name=root, class=org.apache.logging.log4j.core.config.LoggerConfig$RootLogger]. 2025-01-08 04:57:58,869 main DEBUG LoggerConfig$RootLogger$Builder(additivity="null", level="null", levelAndRefs="INFO,Console", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 04:57:58,872 main DEBUG Building Plugin[name=loggers, class=org.apache.logging.log4j.core.config.LoggersPlugin]. 2025-01-08 04:57:58,874 main DEBUG createLoggers(={org.apache.hadoop.metrics2.util.MBeans, org.apache.hadoop.hbase.logging.TestJul2Slf4j, org.apache.zookeeper, org.apache.hadoop.metrics2.impl.MetricsSinkAdapter, org.apache.hadoop.metrics2.impl.MetricsSystemImpl, org.apache.directory, org.apache.hadoop.hbase.ipc.FailedServers, org.apache.hadoop.metrics2.impl.MetricsConfig, org.apache.hadoop.hbase.ScheduledChore, org.apache.hadoop.hbase.regionserver.RSRpcServices, org.apache.hadoop, org.apache.hadoop.hbase, org.apache.hbase.thirdparty.io.netty.channel, root}) 2025-01-08 04:57:58,875 main DEBUG Building Plugin[name=layout, class=org.apache.logging.log4j.core.layout.PatternLayout]. 2025-01-08 04:57:58,879 main DEBUG PatternLayout$Builder(pattern="%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n", PatternSelector=null, Configuration(PropertiesConfig), Replace=null, charset="null", alwaysWriteExceptions="null", disableAnsi="null", noConsoleNoAnsi="null", header="null", footer="null") 2025-01-08 04:57:58,880 main DEBUG PluginManager 'Converter' found 47 plugins 2025-01-08 04:57:58,892 main DEBUG Building Plugin[name=appender, class=org.apache.hadoop.hbase.logging.HBaseTestAppender]. 2025-01-08 04:57:58,898 main DEBUG HBaseTestAppender$Builder(target="SYSTEM_ERR", maxSize="1G", bufferedIo="null", bufferSize="null", immediateFlush="null", ignoreExceptions="null", PatternLayout(%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n), name="Console", Configuration(PropertiesConfig), Filter=null, ={}) 2025-01-08 04:57:58,900 main DEBUG Starting HBaseTestOutputStreamManager SYSTEM_ERR 2025-01-08 04:57:58,901 main DEBUG Building Plugin[name=appenders, class=org.apache.logging.log4j.core.config.AppendersPlugin]. 2025-01-08 04:57:58,901 main DEBUG createAppenders(={Console}) 2025-01-08 04:57:58,903 main DEBUG Configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@4b8d604b initialized 2025-01-08 04:57:58,903 main DEBUG Starting configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@4b8d604b 2025-01-08 04:57:58,904 main DEBUG Started configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@4b8d604b OK. 2025-01-08 04:57:58,904 main DEBUG Shutting down OutputStreamManager SYSTEM_OUT.false.false-1 2025-01-08 04:57:58,905 main DEBUG OutputStream closed 2025-01-08 04:57:58,905 main DEBUG Shut down OutputStreamManager SYSTEM_OUT.false.false-1, all resources released: true 2025-01-08 04:57:58,905 main DEBUG Appender DefaultConsole-1 stopped with status true 2025-01-08 04:57:58,905 main DEBUG Stopped org.apache.logging.log4j.core.config.DefaultConfiguration@37d4349f OK 2025-01-08 04:57:59,087 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6 2025-01-08 04:57:59,090 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=StatusLogger 2025-01-08 04:57:59,092 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=ContextSelector 2025-01-08 04:57:59,094 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name= 2025-01-08 04:57:59,094 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.directory 2025-01-08 04:57:59,095 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSinkAdapter 2025-01-08 04:57:59,095 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.zookeeper 2025-01-08 04:57:59,096 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.logging.TestJul2Slf4j 2025-01-08 04:57:59,097 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSystemImpl 2025-01-08 04:57:59,098 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.util.MBeans 2025-01-08 04:57:59,098 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase 2025-01-08 04:57:59,098 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop 2025-01-08 04:57:59,099 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ipc.FailedServers 2025-01-08 04:57:59,099 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.regionserver.RSRpcServices 2025-01-08 04:57:59,100 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsConfig 2025-01-08 04:57:59,100 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hbase.thirdparty.io.netty.channel 2025-01-08 04:57:59,100 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ScheduledChore 2025-01-08 04:57:59,102 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Appenders,name=Console 2025-01-08 04:57:59,105 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2025-01-08 04:57:59,105 main DEBUG Reconfiguration complete for context[name=1dbd16a6] at URI jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-logging/target/hbase-logging-2.7.0-SNAPSHOT-tests.jar!/log4j2.properties (org.apache.logging.log4j.core.LoggerContext@5812f68b) with optional ClassLoader: null 2025-01-08 04:57:59,105 main DEBUG Shutdown hook enabled. Registering a new one. 2025-01-08 04:57:59,107 main DEBUG LoggerContext[name=1dbd16a6, org.apache.logging.log4j.core.LoggerContext@5812f68b] started OK. 2025-01-08T04:57:59,130 INFO [main {}] hbase.HBaseClassTestRule(94): Test class org.apache.hadoop.hbase.snapshot.TestExportSnapshot timeout: 13 mins 2025-01-08 04:57:59,136 main DEBUG AsyncLogger.ThreadNameStrategy=UNCACHED (user specified null, default is UNCACHED) 2025-01-08 04:57:59,136 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2025-01-08T04:57:59,622 DEBUG [main {}] hbase.HBaseTestingUtility(348): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d 2025-01-08T04:57:59,623 INFO [main {}] hbase.HBaseClassTestRule(94): Test class org.apache.hadoop.hbase.snapshot.TestMobExportSnapshot timeout: 13 mins 2025-01-08T04:57:59,627 INFO [main {}] hbase.HBaseClassTestRule(94): Test class org.apache.hadoop.hbase.snapshot.TestMobSecureExportSnapshot timeout: 13 mins 2025-01-08T04:57:59,729 WARN [Time-limited test {}] util.NativeCodeLoader(60): Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2025-01-08T04:58:00,011 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1126): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2025-01-08T04:58:00,012 INFO [Time-limited test {}] hbase.HBaseZKTestingUtility(82): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab, deleteOnExit=true 2025-01-08T04:58:00,013 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1140): STARTING DFS 2025-01-08T04:58:00,013 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/test.cache.data in system properties and HBase conf 2025-01-08T04:58:00,014 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop.tmp.dir in system properties and HBase conf 2025-01-08T04:58:00,015 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop.log.dir in system properties and HBase conf 2025-01-08T04:58:00,079 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/mapreduce.cluster.local.dir in system properties and HBase conf 2025-01-08T04:58:00,080 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/mapreduce.cluster.temp.dir in system properties and HBase conf 2025-01-08T04:58:00,081 INFO [Time-limited test {}] hbase.HBaseTestingUtility(811): read short circuit is OFF 2025-01-08T04:58:00,189 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2025-01-08T04:58:00,192 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2025-01-08T04:58:00,193 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2025-01-08T04:58:00,194 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/yarn.nodemanager.log-dirs in system properties and HBase conf 2025-01-08T04:58:00,194 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2025-01-08T04:58:00,195 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2025-01-08T04:58:00,196 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2025-01-08T04:58:00,196 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2025-01-08T04:58:00,197 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/dfs.journalnode.edits.dir in system properties and HBase conf 2025-01-08T04:58:00,198 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2025-01-08T04:58:00,198 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/nfs.dump.dir in system properties and HBase conf 2025-01-08T04:58:00,199 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/java.io.tmpdir in system properties and HBase conf 2025-01-08T04:58:00,199 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/dfs.journalnode.edits.dir in system properties and HBase conf 2025-01-08T04:58:00,200 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2025-01-08T04:58:00,200 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2025-01-08T04:58:02,093 WARN [Time-limited test {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties 2025-01-08T04:58:02,410 INFO [Time-limited test {}] log.Log(170): Logging initialized @5626ms to org.eclipse.jetty.util.log.Slf4jLog 2025-01-08T04:58:02,537 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T04:58:02,642 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2025-01-08T04:58:02,731 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2025-01-08T04:58:02,731 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2025-01-08T04:58:02,733 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2025-01-08T04:58:02,787 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T04:58:02,819 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@19c1c4e1{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop.log.dir/,AVAILABLE} 2025-01-08T04:58:02,820 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@15c67ac2{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2025-01-08T04:58:03,231 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@27047d40{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/java.io.tmpdir/jetty-localhost-40771-hadoop-hdfs-3_4_1-tests_jar-_-any-11156743003774098406/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2025-01-08T04:58:03,241 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@61d1e83b{HTTP/1.1, (http/1.1)}{localhost:40771} 2025-01-08T04:58:03,241 INFO [Time-limited test {}] server.Server(415): Started @6458ms 2025-01-08T04:58:03,744 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T04:58:03,753 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2025-01-08T04:58:03,755 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2025-01-08T04:58:03,755 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2025-01-08T04:58:03,755 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2025-01-08T04:58:03,756 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@5ce17cd2{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop.log.dir/,AVAILABLE} 2025-01-08T04:58:03,757 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3c53997b{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2025-01-08T04:58:03,910 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@d799fa8{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/java.io.tmpdir/jetty-localhost-33939-hadoop-hdfs-3_4_1-tests_jar-_-any-2411870409370580492/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2025-01-08T04:58:03,911 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@7455429{HTTP/1.1, (http/1.1)}{localhost:33939} 2025-01-08T04:58:03,911 INFO [Time-limited test {}] server.Server(415): Started @7128ms 2025-01-08T04:58:03,995 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2025-01-08T04:58:04,211 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T04:58:04,218 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2025-01-08T04:58:04,242 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2025-01-08T04:58:04,242 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2025-01-08T04:58:04,243 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2025-01-08T04:58:04,249 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@4cbe4b5e{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop.log.dir/,AVAILABLE} 2025-01-08T04:58:04,250 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@53b7d93e{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2025-01-08T04:58:04,422 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@13469af5{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/java.io.tmpdir/jetty-localhost-39721-hadoop-hdfs-3_4_1-tests_jar-_-any-5962989107489614738/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2025-01-08T04:58:04,424 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@673672ee{HTTP/1.1, (http/1.1)}{localhost:39721} 2025-01-08T04:58:04,425 INFO [Time-limited test {}] server.Server(415): Started @7642ms 2025-01-08T04:58:04,428 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2025-01-08T04:58:04,568 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T04:58:04,576 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2025-01-08T04:58:04,591 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2025-01-08T04:58:04,591 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2025-01-08T04:58:04,591 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2025-01-08T04:58:04,601 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@2f88ae23{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop.log.dir/,AVAILABLE} 2025-01-08T04:58:04,602 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3564bcd4{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2025-01-08T04:58:04,729 WARN [Thread-105 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data1/current/BP-1446750426-172.17.0.3-1736312281490/current, will proceed with Du for space computation calculation, 2025-01-08T04:58:04,730 WARN [Thread-106 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data3/current/BP-1446750426-172.17.0.3-1736312281490/current, will proceed with Du for space computation calculation, 2025-01-08T04:58:04,734 WARN [Thread-108 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data4/current/BP-1446750426-172.17.0.3-1736312281490/current, will proceed with Du for space computation calculation, 2025-01-08T04:58:04,735 WARN [Thread-107 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data2/current/BP-1446750426-172.17.0.3-1736312281490/current, will proceed with Du for space computation calculation, 2025-01-08T04:58:04,828 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@85b88e0{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/java.io.tmpdir/jetty-localhost-40719-hadoop-hdfs-3_4_1-tests_jar-_-any-14959064819910900181/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2025-01-08T04:58:04,831 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@22e0a731{HTTP/1.1, (http/1.1)}{localhost:40719} 2025-01-08T04:58:04,832 INFO [Time-limited test {}] server.Server(415): Started @8049ms 2025-01-08T04:58:04,837 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2025-01-08T04:58:04,881 WARN [Thread-83 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2025-01-08T04:58:04,888 WARN [Thread-58 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2025-01-08T04:58:05,020 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xb04716bfbc57dccb with lease ID 0x7325c11bf735d641: Processing first storage report for DS-0a2953dd-a4c6-4650-9631-02981d9d1e8d from datanode DatanodeRegistration(127.0.0.1:40353, datanodeUuid=952bc167-5e99-4eb6-aae6-febac4c94286, infoPort=33789, infoSecurePort=0, ipcPort=45943, storageInfo=lv=-57;cid=testClusterID;nsid=1408801277;c=1736312281490) 2025-01-08T04:58:05,021 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xb04716bfbc57dccb with lease ID 0x7325c11bf735d641: from storage DS-0a2953dd-a4c6-4650-9631-02981d9d1e8d node DatanodeRegistration(127.0.0.1:40353, datanodeUuid=952bc167-5e99-4eb6-aae6-febac4c94286, infoPort=33789, infoSecurePort=0, ipcPort=45943, storageInfo=lv=-57;cid=testClusterID;nsid=1408801277;c=1736312281490), blocks: 0, hasStaleStorage: true, processing time: 2 msecs, invalidatedBlocks: 0 2025-01-08T04:58:05,022 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x5a35444ca0666956 with lease ID 0x7325c11bf735d642: Processing first storage report for DS-f676faff-d67d-404f-bb0f-b90001fc3639 from datanode DatanodeRegistration(127.0.0.1:33991, datanodeUuid=e8db6b00-be0c-4265-bbbd-8f000659b075, infoPort=45387, infoSecurePort=0, ipcPort=36851, storageInfo=lv=-57;cid=testClusterID;nsid=1408801277;c=1736312281490) 2025-01-08T04:58:05,022 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x5a35444ca0666956 with lease ID 0x7325c11bf735d642: from storage DS-f676faff-d67d-404f-bb0f-b90001fc3639 node DatanodeRegistration(127.0.0.1:33991, datanodeUuid=e8db6b00-be0c-4265-bbbd-8f000659b075, infoPort=45387, infoSecurePort=0, ipcPort=36851, storageInfo=lv=-57;cid=testClusterID;nsid=1408801277;c=1736312281490), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2025-01-08T04:58:05,023 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xb04716bfbc57dccb with lease ID 0x7325c11bf735d641: Processing first storage report for DS-ad9b8c13-7143-49e1-8cf4-fd5e9929460b from datanode DatanodeRegistration(127.0.0.1:40353, datanodeUuid=952bc167-5e99-4eb6-aae6-febac4c94286, infoPort=33789, infoSecurePort=0, ipcPort=45943, storageInfo=lv=-57;cid=testClusterID;nsid=1408801277;c=1736312281490) 2025-01-08T04:58:05,023 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xb04716bfbc57dccb with lease ID 0x7325c11bf735d641: from storage DS-ad9b8c13-7143-49e1-8cf4-fd5e9929460b node DatanodeRegistration(127.0.0.1:40353, datanodeUuid=952bc167-5e99-4eb6-aae6-febac4c94286, infoPort=33789, infoSecurePort=0, ipcPort=45943, storageInfo=lv=-57;cid=testClusterID;nsid=1408801277;c=1736312281490), blocks: 0, hasStaleStorage: false, processing time: 1 msecs, invalidatedBlocks: 0 2025-01-08T04:58:05,023 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x5a35444ca0666956 with lease ID 0x7325c11bf735d642: Processing first storage report for DS-4f64776a-49f1-47ae-b875-f7a9908a9e17 from datanode DatanodeRegistration(127.0.0.1:33991, datanodeUuid=e8db6b00-be0c-4265-bbbd-8f000659b075, infoPort=45387, infoSecurePort=0, ipcPort=36851, storageInfo=lv=-57;cid=testClusterID;nsid=1408801277;c=1736312281490) 2025-01-08T04:58:05,023 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x5a35444ca0666956 with lease ID 0x7325c11bf735d642: from storage DS-4f64776a-49f1-47ae-b875-f7a9908a9e17 node DatanodeRegistration(127.0.0.1:33991, datanodeUuid=e8db6b00-be0c-4265-bbbd-8f000659b075, infoPort=45387, infoSecurePort=0, ipcPort=36851, storageInfo=lv=-57;cid=testClusterID;nsid=1408801277;c=1736312281490), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2025-01-08T04:58:05,192 WARN [Thread-139 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data5/current/BP-1446750426-172.17.0.3-1736312281490/current, will proceed with Du for space computation calculation, 2025-01-08T04:58:05,251 WARN [Thread-140 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data6/current/BP-1446750426-172.17.0.3-1736312281490/current, will proceed with Du for space computation calculation, 2025-01-08T04:58:05,405 WARN [Thread-129 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2025-01-08T04:58:05,428 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xfe7ac372c4eb7b0 with lease ID 0x7325c11bf735d643: Processing first storage report for DS-84b09c6f-eff8-44d2-9077-117e68175eb1 from datanode DatanodeRegistration(127.0.0.1:45867, datanodeUuid=ec76c1ae-a7ea-45af-8315-9b1b87ea684c, infoPort=34673, infoSecurePort=0, ipcPort=38935, storageInfo=lv=-57;cid=testClusterID;nsid=1408801277;c=1736312281490) 2025-01-08T04:58:05,428 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xfe7ac372c4eb7b0 with lease ID 0x7325c11bf735d643: from storage DS-84b09c6f-eff8-44d2-9077-117e68175eb1 node DatanodeRegistration(127.0.0.1:45867, datanodeUuid=ec76c1ae-a7ea-45af-8315-9b1b87ea684c, infoPort=34673, infoSecurePort=0, ipcPort=38935, storageInfo=lv=-57;cid=testClusterID;nsid=1408801277;c=1736312281490), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2025-01-08T04:58:05,428 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xfe7ac372c4eb7b0 with lease ID 0x7325c11bf735d643: Processing first storage report for DS-4adfb765-b337-4ae6-9015-28c5b85c03fd from datanode DatanodeRegistration(127.0.0.1:45867, datanodeUuid=ec76c1ae-a7ea-45af-8315-9b1b87ea684c, infoPort=34673, infoSecurePort=0, ipcPort=38935, storageInfo=lv=-57;cid=testClusterID;nsid=1408801277;c=1736312281490) 2025-01-08T04:58:05,429 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xfe7ac372c4eb7b0 with lease ID 0x7325c11bf735d643: from storage DS-4adfb765-b337-4ae6-9015-28c5b85c03fd node DatanodeRegistration(127.0.0.1:45867, datanodeUuid=ec76c1ae-a7ea-45af-8315-9b1b87ea684c, infoPort=34673, infoSecurePort=0, ipcPort=38935, storageInfo=lv=-57;cid=testClusterID;nsid=1408801277;c=1736312281490), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2025-01-08T04:58:05,777 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(703): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d 2025-01-08T04:58:05,992 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(259): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/zookeeper_0, clientPort=59048, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2025-01-08T04:58:06,015 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(284): Started MiniZooKeeperCluster and ran 'stat' on client port=59048 2025-01-08T04:58:06,040 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T04:58:06,052 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T04:58:06,646 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741825_1001 (size=7) 2025-01-08T04:58:06,648 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741825_1001 (size=7) 2025-01-08T04:58:06,648 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741825_1001 (size=7) 2025-01-08T04:58:07,071 INFO [Time-limited test {}] util.FSUtils(490): Created version file at hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 with version=8 2025-01-08T04:58:07,072 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1460): Setting hbase.fs.tmp.dir to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/hbase-staging 2025-01-08T04:58:07,289 DEBUG [Time-limited test {}] channel.MultithreadEventLoopGroup(44): -Dio.netty.eventLoopThreads: 16 2025-01-08T04:58:07,780 INFO [Time-limited test {}] client.ConnectionUtils(129): master/f28094b72178:0 server-side Connection retries=45 2025-01-08T04:58:07,813 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2025-01-08T04:58:07,814 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2025-01-08T04:58:07,816 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2025-01-08T04:58:07,816 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2025-01-08T04:58:07,816 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2025-01-08T04:58:08,161 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2025-01-08T04:58:08,367 INFO [Time-limited test {}] metrics.MetricRegistriesLoader(60): Loaded MetricRegistries class org.apache.hadoop.hbase.metrics.impl.MetricRegistriesImpl 2025-01-08T04:58:08,385 DEBUG [Time-limited test {}] util.ClassSize(228): Using Unsafe to estimate memory layout 2025-01-08T04:58:08,395 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2025-01-08T04:58:08,442 DEBUG [Time-limited test {}] channel.DefaultChannelId(84): -Dio.netty.processId: 590 (auto-detected) 2025-01-08T04:58:08,443 DEBUG [Time-limited test {}] channel.DefaultChannelId(106): -Dio.netty.machineId: 02:42:ac:ff:fe:11:00:03 (auto-detected) 2025-01-08T04:58:08,483 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.3:42195 2025-01-08T04:58:08,501 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T04:58:08,506 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T04:58:08,532 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=master:42195 connecting to ZooKeeper ensemble=127.0.0.1:59048 2025-01-08T04:58:08,581 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:421950x0, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2025-01-08T04:58:08,596 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:42195-0x1000ae1399f0000 connected 2025-01-08T04:58:08,645 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2025-01-08T04:58:08,648 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2025-01-08T04:58:08,666 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2025-01-08T04:58:08,673 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=42195 2025-01-08T04:58:08,674 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=42195 2025-01-08T04:58:08,674 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=42195 2025-01-08T04:58:08,675 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=42195 2025-01-08T04:58:08,676 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=42195 2025-01-08T04:58:08,686 INFO [Time-limited test {}] master.HMaster(499): hbase.rootdir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49, hbase.cluster.distributed=false 2025-01-08T04:58:08,780 INFO [Time-limited test {}] client.ConnectionUtils(129): regionserver/f28094b72178:0 server-side Connection retries=45 2025-01-08T04:58:08,780 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2025-01-08T04:58:08,780 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2025-01-08T04:58:08,780 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2025-01-08T04:58:08,781 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2025-01-08T04:58:08,781 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2025-01-08T04:58:08,783 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2025-01-08T04:58:08,786 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2025-01-08T04:58:08,794 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.3:39801 2025-01-08T04:58:08,796 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2025-01-08T04:58:08,806 INFO [Time-limited test {}] mob.MobFileCache(128): MobFileCache disabled 2025-01-08T04:58:08,808 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T04:58:08,811 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T04:58:08,815 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=regionserver:39801 connecting to ZooKeeper ensemble=127.0.0.1:59048 2025-01-08T04:58:08,828 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:398010x0, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2025-01-08T04:58:08,829 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:39801-0x1000ae1399f0001 connected 2025-01-08T04:58:08,830 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2025-01-08T04:58:08,837 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2025-01-08T04:58:08,840 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2025-01-08T04:58:08,843 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=39801 2025-01-08T04:58:08,845 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=39801 2025-01-08T04:58:08,848 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=39801 2025-01-08T04:58:08,850 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=39801 2025-01-08T04:58:08,851 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=39801 2025-01-08T04:58:08,874 INFO [Time-limited test {}] client.ConnectionUtils(129): regionserver/f28094b72178:0 server-side Connection retries=45 2025-01-08T04:58:08,875 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2025-01-08T04:58:08,875 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2025-01-08T04:58:08,875 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2025-01-08T04:58:08,875 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2025-01-08T04:58:08,876 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2025-01-08T04:58:08,876 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2025-01-08T04:58:08,876 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2025-01-08T04:58:08,878 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.3:37583 2025-01-08T04:58:08,879 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2025-01-08T04:58:08,881 INFO [Time-limited test {}] mob.MobFileCache(128): MobFileCache disabled 2025-01-08T04:58:08,882 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T04:58:08,885 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T04:58:08,892 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=regionserver:37583 connecting to ZooKeeper ensemble=127.0.0.1:59048 2025-01-08T04:58:08,901 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:375830x0, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2025-01-08T04:58:08,901 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:375830x0, quorum=127.0.0.1:59048, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2025-01-08T04:58:08,903 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:37583-0x1000ae1399f0002 connected 2025-01-08T04:58:08,904 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2025-01-08T04:58:08,907 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2025-01-08T04:58:08,909 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=37583 2025-01-08T04:58:08,910 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=37583 2025-01-08T04:58:08,913 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=37583 2025-01-08T04:58:08,914 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=37583 2025-01-08T04:58:08,915 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=37583 2025-01-08T04:58:08,943 INFO [Time-limited test {}] client.ConnectionUtils(129): regionserver/f28094b72178:0 server-side Connection retries=45 2025-01-08T04:58:08,943 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2025-01-08T04:58:08,944 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2025-01-08T04:58:08,944 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2025-01-08T04:58:08,944 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2025-01-08T04:58:08,944 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2025-01-08T04:58:08,944 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2025-01-08T04:58:08,945 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2025-01-08T04:58:08,946 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.3:37967 2025-01-08T04:58:08,947 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2025-01-08T04:58:08,948 INFO [Time-limited test {}] mob.MobFileCache(128): MobFileCache disabled 2025-01-08T04:58:08,950 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T04:58:08,954 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T04:58:08,959 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=regionserver:37967 connecting to ZooKeeper ensemble=127.0.0.1:59048 2025-01-08T04:58:08,964 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:379670x0, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2025-01-08T04:58:08,966 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:379670x0, quorum=127.0.0.1:59048, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2025-01-08T04:58:08,966 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:37967-0x1000ae1399f0003 connected 2025-01-08T04:58:08,968 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2025-01-08T04:58:08,969 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2025-01-08T04:58:08,969 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=37967 2025-01-08T04:58:08,970 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=37967 2025-01-08T04:58:08,970 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=37967 2025-01-08T04:58:08,971 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=37967 2025-01-08T04:58:08,971 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=37967 2025-01-08T04:58:08,976 INFO [master/f28094b72178:0:becomeActiveMaster {}] master.HMaster(2460): Adding backup master ZNode /hbase/backup-masters/f28094b72178,42195,1736312287279 2025-01-08T04:58:08,983 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2025-01-08T04:58:08,983 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2025-01-08T04:58:08,983 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2025-01-08T04:58:08,983 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2025-01-08T04:58:08,988 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/f28094b72178,42195,1736312287279 2025-01-08T04:58:08,996 DEBUG [M:0;f28094b72178:42195 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;f28094b72178:42195 2025-01-08T04:58:09,014 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2025-01-08T04:58:09,015 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:09,015 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2025-01-08T04:58:09,015 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2025-01-08T04:58:09,015 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:09,015 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:09,015 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2025-01-08T04:58:09,016 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:09,018 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2025-01-08T04:58:09,019 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(111): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2025-01-08T04:58:09,022 INFO [master/f28094b72178:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/f28094b72178,42195,1736312287279 from backup master directory 2025-01-08T04:58:09,026 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2025-01-08T04:58:09,026 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2025-01-08T04:58:09,026 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2025-01-08T04:58:09,026 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/f28094b72178,42195,1736312287279 2025-01-08T04:58:09,027 WARN [master/f28094b72178:0:becomeActiveMaster {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2025-01-08T04:58:09,027 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2025-01-08T04:58:09,027 INFO [master/f28094b72178:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=f28094b72178,42195,1736312287279 2025-01-08T04:58:09,031 INFO [master/f28094b72178:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating data MemStoreChunkPool with chunk size 2 MB, max count 396, initial count 0 2025-01-08T04:58:09,033 INFO [master/f28094b72178:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating index MemStoreChunkPool with chunk size 204.80 KB, max count 440, initial count 0 2025-01-08T04:58:09,204 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741826_1002 (size=42) 2025-01-08T04:58:09,204 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741826_1002 (size=42) 2025-01-08T04:58:09,205 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741826_1002 (size=42) 2025-01-08T04:58:09,211 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] util.FSUtils(639): Created cluster ID file at hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/hbase.id with ID: 4defa70b-5188-4ebc-a40b-96cf0c2825fc 2025-01-08T04:58:09,274 INFO [master/f28094b72178:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T04:58:09,313 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:09,313 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:09,314 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:09,314 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:09,393 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741827_1003 (size=196) 2025-01-08T04:58:09,395 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741827_1003 (size=196) 2025-01-08T04:58:09,396 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741827_1003 (size=196) 2025-01-08T04:58:09,423 INFO [master/f28094b72178:0:becomeActiveMaster {}] region.MasterRegion(372): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T04:58:09,425 INFO [master/f28094b72178:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2025-01-08T04:58:09,454 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(244): No decryptEncryptedDataEncryptionKey method in DFSClient, should be hadoop version with HDFS-12396 java.lang.NoSuchMethodException: org.apache.hadoop.hdfs.DFSClient.decryptEncryptedDataEncryptionKey(org.apache.hadoop.fs.FileEncryptionInfo) at java.lang.Class.getDeclaredMethod(Class.java:2675) ~[?:?] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelperWithoutHDFS12396(FanOutOneBlockAsyncDFSOutputSaslHelper.java:183) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelper(FanOutOneBlockAsyncDFSOutputSaslHelper.java:242) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.(FanOutOneBlockAsyncDFSOutputSaslHelper.java:253) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at java.lang.Class.forName0(Native Method) ~[?:?] at java.lang.Class.forName(Class.java:375) ~[?:?] at org.apache.hadoop.hbase.wal.AsyncFSWALProvider.load(AsyncFSWALProvider.java:147) ~[classes/:?] at org.apache.hadoop.hbase.wal.WALFactory.getProviderClass(WALFactory.java:160) ~[classes/:?] at org.apache.hadoop.hbase.wal.WALFactory.getProvider(WALFactory.java:200) ~[classes/:?] at org.apache.hadoop.hbase.wal.WALFactory.(WALFactory.java:232) ~[classes/:?] at org.apache.hadoop.hbase.wal.WALFactory.(WALFactory.java:207) ~[classes/:?] at org.apache.hadoop.hbase.master.region.MasterRegion.create(MasterRegion.java:402) ~[classes/:?] at org.apache.hadoop.hbase.master.region.MasterRegionFactory.create(MasterRegionFactory.java:135) ~[classes/:?] at org.apache.hadoop.hbase.master.HMaster.finishActiveMasterInitialization(HMaster.java:984) ~[classes/:?] at org.apache.hadoop.hbase.master.HMaster.startActiveMasterManager(HMaster.java:2485) ~[classes/:?] at org.apache.hadoop.hbase.master.HMaster.lambda$run$0(HMaster.java:601) ~[classes/:?] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) ~[hbase-common-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) ~[hbase-common-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.master.HMaster.lambda$run$1(HMaster.java:598) ~[classes/:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2025-01-08T04:58:09,467 INFO [master/f28094b72178:0:becomeActiveMaster {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2025-01-08T04:58:09,592 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741828_1004 (size=1189) 2025-01-08T04:58:09,603 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741828_1004 (size=1189) 2025-01-08T04:58:09,604 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741828_1004 (size=1189) 2025-01-08T04:58:09,639 INFO [master/f28094b72178:0:becomeActiveMaster {}] regionserver.HRegion(7124): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/MasterData/data/master/store 2025-01-08T04:58:09,754 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741829_1005 (size=34) 2025-01-08T04:58:09,757 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741829_1005 (size=34) 2025-01-08T04:58:09,757 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741829_1005 (size=34) 2025-01-08T04:58:09,768 INFO [master/f28094b72178:0:becomeActiveMaster {}] throttle.StoreHotnessProtector(112): StoreHotnessProtector is disabled. Set hbase.region.store.parallel.put.limit > 0 to enable, which may help mitigate load under heavy write pressure. 2025-01-08T04:58:09,769 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T04:58:09,771 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2025-01-08T04:58:09,771 INFO [master/f28094b72178:0:becomeActiveMaster {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2025-01-08T04:58:09,772 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2025-01-08T04:58:09,772 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2025-01-08T04:58:09,772 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2025-01-08T04:58:09,772 INFO [master/f28094b72178:0:becomeActiveMaster {}] regionserver.HRegion(1922): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2025-01-08T04:58:09,772 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2025-01-08T04:58:09,779 WARN [master/f28094b72178:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/MasterData/data/master/store/.initializing 2025-01-08T04:58:09,779 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/MasterData/WALs/f28094b72178,42195,1736312287279 2025-01-08T04:58:09,797 INFO [master/f28094b72178:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2025-01-08T04:58:09,836 INFO [master/f28094b72178:0:becomeActiveMaster {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=f28094b72178%2C42195%2C1736312287279, suffix=, logDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/MasterData/WALs/f28094b72178,42195,1736312287279, archiveDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/MasterData/oldWALs, maxLogs=10 2025-01-08T04:58:09,909 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(617): When create output stream for /user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/MasterData/WALs/f28094b72178,42195,1736312287279/f28094b72178%2C42195%2C1736312287279.1736312289851, exclude list is [], retry=0 2025-01-08T04:58:09,949 DEBUG [RS-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:33991,DS-f676faff-d67d-404f-bb0f-b90001fc3639,DISK] 2025-01-08T04:58:09,950 DEBUG [RS-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:40353,DS-0a2953dd-a4c6-4650-9631-02981d9d1e8d,DISK] 2025-01-08T04:58:09,951 DEBUG [RS-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45867,DS-84b09c6f-eff8-44d2-9077-117e68175eb1,DISK] 2025-01-08T04:58:09,954 DEBUG [RS-EventLoopGroup-5-1 {}] asyncfs.ProtobufDecoder(117): Hadoop 3.3 and above shades protobuf. 2025-01-08T04:58:10,034 INFO [master/f28094b72178:0:becomeActiveMaster {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/MasterData/WALs/f28094b72178,42195,1736312287279/f28094b72178%2C42195%2C1736312287279.1736312289851 2025-01-08T04:58:10,036 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] wal.AbstractFSWAL(925): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:45387:45387),(127.0.0.1/127.0.0.1:34673:34673),(127.0.0.1/127.0.0.1:33789:33789)] 2025-01-08T04:58:10,036 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] regionserver.HRegion(7285): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2025-01-08T04:58:10,037 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T04:58:10,042 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] regionserver.HRegion(7327): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2025-01-08T04:58:10,044 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] regionserver.HRegion(7330): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2025-01-08T04:58:10,134 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2025-01-08T04:58:10,174 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2025-01-08T04:58:10,180 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:58:10,184 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2025-01-08T04:58:10,184 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2025-01-08T04:58:10,197 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2025-01-08T04:58:10,198 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:58:10,199 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T04:58:10,200 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2025-01-08T04:58:10,203 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2025-01-08T04:58:10,203 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:58:10,207 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T04:58:10,208 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2025-01-08T04:58:10,212 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2025-01-08T04:58:10,212 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:58:10,214 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T04:58:10,220 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2025-01-08T04:58:10,221 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2025-01-08T04:58:10,235 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2025-01-08T04:58:10,245 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] regionserver.HRegion(1085): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2025-01-08T04:58:10,258 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T04:58:10,260 INFO [master/f28094b72178:0:becomeActiveMaster {}] regionserver.HRegion(1102): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=64638164, jitterRate=-0.036816298961639404}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2025-01-08T04:58:10,266 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] regionserver.HRegion(1001): Region open journal for 1595e783b53d99cd5eef43b6debb2682: 2025-01-08T04:58:10,271 INFO [master/f28094b72178:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2025-01-08T04:58:10,318 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@3793b353, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T04:58:10,371 INFO [master/f28094b72178:0:becomeActiveMaster {}] master.HMaster(893): No meta location available on zookeeper, skip migrating... 2025-01-08T04:58:10,386 INFO [master/f28094b72178:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2025-01-08T04:58:10,386 INFO [master/f28094b72178:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(633): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2025-01-08T04:58:10,389 INFO [master/f28094b72178:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2025-01-08T04:58:10,391 INFO [master/f28094b72178:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(653): Recovered RegionProcedureStore lease in 1 msec 2025-01-08T04:58:10,398 INFO [master/f28094b72178:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(667): Loaded RegionProcedureStore in 6 msec 2025-01-08T04:58:10,398 INFO [master/f28094b72178:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2025-01-08T04:58:10,435 INFO [master/f28094b72178:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2025-01-08T04:58:10,469 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2025-01-08T04:58:10,474 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/balancer already deleted, retry=false 2025-01-08T04:58:10,481 INFO [master/f28094b72178:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2025-01-08T04:58:10,489 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2025-01-08T04:58:10,494 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/normalizer already deleted, retry=false 2025-01-08T04:58:10,497 INFO [master/f28094b72178:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2025-01-08T04:58:10,504 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2025-01-08T04:58:10,508 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/split already deleted, retry=false 2025-01-08T04:58:10,509 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2025-01-08T04:58:10,514 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/merge already deleted, retry=false 2025-01-08T04:58:10,534 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2025-01-08T04:58:10,538 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/snapshot-cleanup already deleted, retry=false 2025-01-08T04:58:10,546 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2025-01-08T04:58:10,546 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2025-01-08T04:58:10,546 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:10,547 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2025-01-08T04:58:10,547 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:10,546 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:10,549 INFO [master/f28094b72178:0:becomeActiveMaster {}] master.HMaster(837): Active/primary master=f28094b72178,42195,1736312287279, sessionid=0x1000ae1399f0000, setting cluster-up flag (Was=false) 2025-01-08T04:58:10,550 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2025-01-08T04:58:10,550 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:10,575 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:10,576 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:10,576 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:10,580 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:10,595 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2025-01-08T04:58:10,602 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=f28094b72178,42195,1736312287279 2025-01-08T04:58:10,611 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:10,612 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:10,612 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:10,613 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:10,627 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2025-01-08T04:58:10,629 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=f28094b72178,42195,1736312287279 2025-01-08T04:58:10,736 DEBUG [RS:0;f28094b72178:39801 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;f28094b72178:39801 2025-01-08T04:58:10,740 INFO [RS:0;f28094b72178:39801 {}] regionserver.HRegionServer(1008): ClusterId : 4defa70b-5188-4ebc-a40b-96cf0c2825fc 2025-01-08T04:58:10,741 DEBUG [RS:1;f28094b72178:37583 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;f28094b72178:37583 2025-01-08T04:58:10,745 DEBUG [RS:0;f28094b72178:39801 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2025-01-08T04:58:10,746 INFO [RS:1;f28094b72178:37583 {}] regionserver.HRegionServer(1008): ClusterId : 4defa70b-5188-4ebc-a40b-96cf0c2825fc 2025-01-08T04:58:10,746 DEBUG [RS:1;f28094b72178:37583 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2025-01-08T04:58:10,755 DEBUG [RS:1;f28094b72178:37583 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2025-01-08T04:58:10,755 DEBUG [RS:1;f28094b72178:37583 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2025-01-08T04:58:10,756 DEBUG [RS:0;f28094b72178:39801 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2025-01-08T04:58:10,756 DEBUG [RS:0;f28094b72178:39801 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2025-01-08T04:58:10,756 DEBUG [RS:2;f28094b72178:37967 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;f28094b72178:37967 2025-01-08T04:58:10,760 DEBUG [RS:0;f28094b72178:39801 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2025-01-08T04:58:10,761 DEBUG [RS:1;f28094b72178:37583 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2025-01-08T04:58:10,762 DEBUG [RS:1;f28094b72178:37583 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1b8090ff, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T04:58:10,761 DEBUG [RS:0;f28094b72178:39801 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@165d904f, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T04:58:10,765 INFO [RS:2;f28094b72178:37967 {}] regionserver.HRegionServer(1008): ClusterId : 4defa70b-5188-4ebc-a40b-96cf0c2825fc 2025-01-08T04:58:10,765 DEBUG [RS:2;f28094b72178:37967 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2025-01-08T04:58:10,770 DEBUG [RS:1;f28094b72178:37583 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@e86ac4b, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=f28094b72178/172.17.0.3:0 2025-01-08T04:58:10,770 DEBUG [RS:0;f28094b72178:39801 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@6e1d9065, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=f28094b72178/172.17.0.3:0 2025-01-08T04:58:10,774 INFO [RS:0;f28094b72178:39801 {}] regionserver.RegionServerCoprocessorHost(67): System coprocessor loading is enabled 2025-01-08T04:58:10,774 DEBUG [RS:2;f28094b72178:37967 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2025-01-08T04:58:10,774 DEBUG [RS:2;f28094b72178:37967 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2025-01-08T04:58:10,774 INFO [RS:0;f28094b72178:39801 {}] regionserver.RegionServerCoprocessorHost(68): Table coprocessor loading is enabled 2025-01-08T04:58:10,775 INFO [RS:1;f28094b72178:37583 {}] regionserver.RegionServerCoprocessorHost(67): System coprocessor loading is enabled 2025-01-08T04:58:10,775 INFO [RS:1;f28094b72178:37583 {}] regionserver.RegionServerCoprocessorHost(68): Table coprocessor loading is enabled 2025-01-08T04:58:10,781 DEBUG [RS:2;f28094b72178:37967 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2025-01-08T04:58:10,782 DEBUG [RS:2;f28094b72178:37967 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2de71e81, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T04:58:10,793 DEBUG [RS:2;f28094b72178:37967 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@396ac8e9, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=f28094b72178/172.17.0.3:0 2025-01-08T04:58:10,793 INFO [RS:2;f28094b72178:37967 {}] regionserver.RegionServerCoprocessorHost(67): System coprocessor loading is enabled 2025-01-08T04:58:10,793 INFO [RS:2;f28094b72178:37967 {}] regionserver.RegionServerCoprocessorHost(68): Table coprocessor loading is enabled 2025-01-08T04:58:10,838 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] master.HMaster(3405): Registered master coprocessor service: service=AccessControlService 2025-01-08T04:58:10,839 DEBUG [RS:1;f28094b72178:37583 {}] regionserver.HRegionServer(900): Registered regionserver coprocessor executorService: executorService=AccessControlService 2025-01-08T04:58:10,839 DEBUG [RS:2;f28094b72178:37967 {}] regionserver.HRegionServer(900): Registered regionserver coprocessor executorService: executorService=AccessControlService 2025-01-08T04:58:10,839 DEBUG [RS:0;f28094b72178:39801 {}] regionserver.HRegionServer(900): Registered regionserver coprocessor executorService: executorService=AccessControlService 2025-01-08T04:58:10,840 INFO [RS:1;f28094b72178:37583 {}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T04:58:10,840 INFO [RS:2;f28094b72178:37967 {}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T04:58:10,840 DEBUG [RS:1;f28094b72178:37583 {}] regionserver.HRegionServer(1090): About to register with Master. 2025-01-08T04:58:10,840 DEBUG [RS:2;f28094b72178:37967 {}] regionserver.HRegionServer(1090): About to register with Master. 2025-01-08T04:58:10,841 INFO [RS:0;f28094b72178:39801 {}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T04:58:10,841 DEBUG [RS:0;f28094b72178:39801 {}] regionserver.HRegionServer(1090): About to register with Master. 2025-01-08T04:58:10,842 INFO [master/f28094b72178:0:becomeActiveMaster {}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T04:58:10,842 INFO [master/f28094b72178:0:becomeActiveMaster {}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.SecureTestUtil$MasterSyncObserver loaded, priority=536870912. 2025-01-08T04:58:10,843 INFO [RS:1;f28094b72178:37583 {}] regionserver.HRegionServer(3073): reportForDuty to master=f28094b72178,42195,1736312287279 with isa=f28094b72178/172.17.0.3:37583, startcode=1736312288873 2025-01-08T04:58:10,844 INFO [RS:0;f28094b72178:39801 {}] regionserver.HRegionServer(3073): reportForDuty to master=f28094b72178,42195,1736312287279 with isa=f28094b72178/172.17.0.3:39801, startcode=1736312288779 2025-01-08T04:58:10,846 INFO [RS:2;f28094b72178:37967 {}] regionserver.HRegionServer(3073): reportForDuty to master=f28094b72178,42195,1736312287279 with isa=f28094b72178/172.17.0.3:37967, startcode=1736312288941 2025-01-08T04:58:10,891 DEBUG [RS:2;f28094b72178:37967 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2025-01-08T04:58:10,891 DEBUG [RS:0;f28094b72178:39801 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2025-01-08T04:58:10,891 DEBUG [RS:1;f28094b72178:37583 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2025-01-08T04:58:10,962 INFO [RS-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:37569, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.1 (auth:SIMPLE), service=RegionServerStatusService 2025-01-08T04:58:10,962 INFO [RS-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:54641, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.2 (auth:SIMPLE), service=RegionServerStatusService 2025-01-08T04:58:10,962 INFO [RS-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:33861, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=RegionServerStatusService 2025-01-08T04:58:10,971 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42195 {}] ipc.MetricsHBaseServer(152): Unknown exception type org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server is not running yet at org.apache.hadoop.hbase.master.HMaster.checkServiceStarted(HMaster.java:3295) ~[classes/:?] at org.apache.hadoop.hbase.master.MasterRpcServices.regionServerStartup(MasterRpcServices.java:593) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$2.callBlockingMethod(RegionServerStatusProtos.java:16714) ~[hbase-protocol-shaded-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:443) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:105) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) ~[classes/:2.7.0-SNAPSHOT] 2025-01-08T04:58:10,974 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT; InitMetaProcedure table=hbase:meta 2025-01-08T04:58:10,987 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42195 {}] ipc.MetricsHBaseServer(152): Unknown exception type org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server is not running yet at org.apache.hadoop.hbase.master.HMaster.checkServiceStarted(HMaster.java:3295) ~[classes/:?] at org.apache.hadoop.hbase.master.MasterRpcServices.regionServerStartup(MasterRpcServices.java:593) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$2.callBlockingMethod(RegionServerStatusProtos.java:16714) ~[hbase-protocol-shaded-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:443) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:105) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) ~[classes/:2.7.0-SNAPSHOT] 2025-01-08T04:58:10,988 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42195 {}] ipc.MetricsHBaseServer(152): Unknown exception type org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server is not running yet at org.apache.hadoop.hbase.master.HMaster.checkServiceStarted(HMaster.java:3295) ~[classes/:?] at org.apache.hadoop.hbase.master.MasterRpcServices.regionServerStartup(MasterRpcServices.java:593) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$2.callBlockingMethod(RegionServerStatusProtos.java:16714) ~[hbase-protocol-shaded-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:443) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:105) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) ~[classes/:2.7.0-SNAPSHOT] 2025-01-08T04:58:10,993 INFO [master/f28094b72178:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(575): slop=0.2 2025-01-08T04:58:10,998 INFO [master/f28094b72178:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(294): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2025-01-08T04:58:11,009 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] balancer.RegionLocationFinder(146): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: f28094b72178,42195,1736312287279 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2025-01-08T04:58:11,014 DEBUG [RS:0;f28094b72178:39801 {}] regionserver.HRegionServer(3097): Master is not running yet 2025-01-08T04:58:11,014 WARN [RS:0;f28094b72178:39801 {}] regionserver.HRegionServer(1099): reportForDuty failed; sleeping 100 ms and then retrying. 2025-01-08T04:58:11,014 DEBUG [RS:2;f28094b72178:37967 {}] regionserver.HRegionServer(3097): Master is not running yet 2025-01-08T04:58:11,015 WARN [RS:2;f28094b72178:37967 {}] regionserver.HRegionServer(1099): reportForDuty failed; sleeping 100 ms and then retrying. 2025-01-08T04:58:11,015 DEBUG [RS:1;f28094b72178:37583 {}] regionserver.HRegionServer(3097): Master is not running yet 2025-01-08T04:58:11,015 WARN [RS:1;f28094b72178:37583 {}] regionserver.HRegionServer(1099): reportForDuty failed; sleeping 100 ms and then retrying. 2025-01-08T04:58:11,016 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/f28094b72178:0, corePoolSize=5, maxPoolSize=5 2025-01-08T04:58:11,016 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/f28094b72178:0, corePoolSize=5, maxPoolSize=5 2025-01-08T04:58:11,016 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/f28094b72178:0, corePoolSize=5, maxPoolSize=5 2025-01-08T04:58:11,016 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/f28094b72178:0, corePoolSize=5, maxPoolSize=5 2025-01-08T04:58:11,017 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/f28094b72178:0, corePoolSize=10, maxPoolSize=10 2025-01-08T04:58:11,017 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,017 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/f28094b72178:0, corePoolSize=2, maxPoolSize=2 2025-01-08T04:58:11,017 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,042 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, locked=true; InitMetaProcedure table=hbase:meta 2025-01-08T04:58:11,042 INFO [master/f28094b72178:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1736312321042 2025-01-08T04:58:11,043 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(75): BOOTSTRAP: creating hbase:meta region 2025-01-08T04:58:11,044 INFO [master/f28094b72178:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2025-01-08T04:58:11,046 INFO [master/f28094b72178:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2025-01-08T04:58:11,048 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:58:11,048 INFO [PEWorker-1 {}] util.FSTableDescriptors(133): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2025-01-08T04:58:11,051 INFO [master/f28094b72178:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2025-01-08T04:58:11,051 INFO [master/f28094b72178:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2025-01-08T04:58:11,052 INFO [master/f28094b72178:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2025-01-08T04:58:11,052 INFO [master/f28094b72178:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2025-01-08T04:58:11,053 INFO [master/f28094b72178:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:11,059 INFO [master/f28094b72178:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2025-01-08T04:58:11,060 INFO [master/f28094b72178:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2025-01-08T04:58:11,061 INFO [master/f28094b72178:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2025-01-08T04:58:11,075 INFO [master/f28094b72178:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2025-01-08T04:58:11,076 INFO [master/f28094b72178:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2025-01-08T04:58:11,093 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/f28094b72178:0:becomeActiveMaster-HFileCleaner.large.0-1736312291078,5,FailOnTimeoutGroup] 2025-01-08T04:58:11,102 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/f28094b72178:0:becomeActiveMaster-HFileCleaner.small.0-1736312291094,5,FailOnTimeoutGroup] 2025-01-08T04:58:11,103 INFO [master/f28094b72178:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:11,103 INFO [master/f28094b72178:0:becomeActiveMaster {}] master.HMaster(1691): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2025-01-08T04:58:11,105 INFO [master/f28094b72178:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:11,105 INFO [master/f28094b72178:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:11,114 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741831_1007 (size=1039) 2025-01-08T04:58:11,116 INFO [RS:2;f28094b72178:37967 {}] regionserver.HRegionServer(3073): reportForDuty to master=f28094b72178,42195,1736312287279 with isa=f28094b72178/172.17.0.3:37967, startcode=1736312288941 2025-01-08T04:58:11,116 INFO [RS:0;f28094b72178:39801 {}] regionserver.HRegionServer(3073): reportForDuty to master=f28094b72178,42195,1736312287279 with isa=f28094b72178/172.17.0.3:39801, startcode=1736312288779 2025-01-08T04:58:11,116 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741831_1007 (size=1039) 2025-01-08T04:58:11,117 INFO [RS:1;f28094b72178:37583 {}] regionserver.HRegionServer(3073): reportForDuty to master=f28094b72178,42195,1736312287279 with isa=f28094b72178/172.17.0.3:37583, startcode=1736312288873 2025-01-08T04:58:11,118 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42195 {}] master.ServerManager(332): Checking decommissioned status of RegionServer f28094b72178,39801,1736312288779 2025-01-08T04:58:11,119 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741831_1007 (size=1039) 2025-01-08T04:58:11,121 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42195 {}] master.ServerManager(486): Registering regionserver=f28094b72178,39801,1736312288779 2025-01-08T04:58:11,122 INFO [PEWorker-1 {}] util.FSTableDescriptors(140): Updated hbase:meta table descriptor to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1039 2025-01-08T04:58:11,122 INFO [PEWorker-1 {}] regionserver.HRegion(7106): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T04:58:11,132 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42195 {}] master.ServerManager(332): Checking decommissioned status of RegionServer f28094b72178,37967,1736312288941 2025-01-08T04:58:11,132 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42195 {}] master.ServerManager(486): Registering regionserver=f28094b72178,37967,1736312288941 2025-01-08T04:58:11,132 DEBUG [RS:0;f28094b72178:39801 {}] regionserver.HRegionServer(1725): Config from master: hbase.rootdir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T04:58:11,132 DEBUG [RS:0;f28094b72178:39801 {}] regionserver.HRegionServer(1725): Config from master: fs.defaultFS=hdfs://localhost:45939 2025-01-08T04:58:11,132 DEBUG [RS:0;f28094b72178:39801 {}] regionserver.HRegionServer(1725): Config from master: hbase.master.info.port=-1 2025-01-08T04:58:11,136 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42195 {}] master.ServerManager(332): Checking decommissioned status of RegionServer f28094b72178,37583,1736312288873 2025-01-08T04:58:11,137 DEBUG [RS:2;f28094b72178:37967 {}] regionserver.HRegionServer(1725): Config from master: hbase.rootdir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T04:58:11,137 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42195 {}] master.ServerManager(486): Registering regionserver=f28094b72178,37583,1736312288873 2025-01-08T04:58:11,137 DEBUG [RS:2;f28094b72178:37967 {}] regionserver.HRegionServer(1725): Config from master: fs.defaultFS=hdfs://localhost:45939 2025-01-08T04:58:11,137 DEBUG [RS:2;f28094b72178:37967 {}] regionserver.HRegionServer(1725): Config from master: hbase.master.info.port=-1 2025-01-08T04:58:11,137 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2025-01-08T04:58:11,140 DEBUG [RS:1;f28094b72178:37583 {}] regionserver.HRegionServer(1725): Config from master: hbase.rootdir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T04:58:11,141 DEBUG [RS:1;f28094b72178:37583 {}] regionserver.HRegionServer(1725): Config from master: fs.defaultFS=hdfs://localhost:45939 2025-01-08T04:58:11,141 DEBUG [RS:1;f28094b72178:37583 {}] regionserver.HRegionServer(1725): Config from master: hbase.master.info.port=-1 2025-01-08T04:58:11,145 DEBUG [RS:0;f28094b72178:39801 {}] zookeeper.ZKUtil(111): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/f28094b72178,39801,1736312288779 2025-01-08T04:58:11,145 WARN [RS:0;f28094b72178:39801 {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2025-01-08T04:58:11,145 INFO [RS:0;f28094b72178:39801 {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2025-01-08T04:58:11,145 DEBUG [RS:0;f28094b72178:39801 {}] regionserver.HRegionServer(2100): logDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/WALs/f28094b72178,39801,1736312288779 2025-01-08T04:58:11,151 DEBUG [RS:1;f28094b72178:37583 {}] zookeeper.ZKUtil(111): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/f28094b72178,37583,1736312288873 2025-01-08T04:58:11,151 DEBUG [RS:2;f28094b72178:37967 {}] zookeeper.ZKUtil(111): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/f28094b72178,37967,1736312288941 2025-01-08T04:58:11,151 WARN [RS:1;f28094b72178:37583 {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2025-01-08T04:58:11,151 WARN [RS:2;f28094b72178:37967 {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2025-01-08T04:58:11,151 INFO [RS:1;f28094b72178:37583 {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2025-01-08T04:58:11,151 INFO [RS:2;f28094b72178:37967 {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2025-01-08T04:58:11,151 DEBUG [RS:1;f28094b72178:37583 {}] regionserver.HRegionServer(2100): logDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/WALs/f28094b72178,37583,1736312288873 2025-01-08T04:58:11,151 DEBUG [RS:2;f28094b72178:37967 {}] regionserver.HRegionServer(2100): logDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/WALs/f28094b72178,37967,1736312288941 2025-01-08T04:58:11,154 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [f28094b72178,39801,1736312288779] 2025-01-08T04:58:11,154 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [f28094b72178,37583,1736312288873] 2025-01-08T04:58:11,154 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [f28094b72178,37967,1736312288941] 2025-01-08T04:58:11,168 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741832_1008 (size=32) 2025-01-08T04:58:11,168 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741832_1008 (size=32) 2025-01-08T04:58:11,168 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741832_1008 (size=32) 2025-01-08T04:58:11,171 DEBUG [PEWorker-1 {}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T04:58:11,174 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2025-01-08T04:58:11,184 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2025-01-08T04:58:11,184 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:58:11,185 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2025-01-08T04:58:11,185 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2025-01-08T04:58:11,188 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2025-01-08T04:58:11,188 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:58:11,189 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2025-01-08T04:58:11,189 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2025-01-08T04:58:11,191 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2025-01-08T04:58:11,191 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:58:11,192 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2025-01-08T04:58:11,194 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/1588230740 2025-01-08T04:58:11,195 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/1588230740 2025-01-08T04:58:11,200 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (42.7 M)) instead. 2025-01-08T04:58:11,203 DEBUG [RS:1;f28094b72178:37583 {}] regionserver.Replication(140): Replication stats-in-log period=300 seconds 2025-01-08T04:58:11,203 DEBUG [RS:0;f28094b72178:39801 {}] regionserver.Replication(140): Replication stats-in-log period=300 seconds 2025-01-08T04:58:11,205 DEBUG [PEWorker-1 {}] regionserver.HRegion(1085): writing seq id for 1588230740 2025-01-08T04:58:11,209 DEBUG [RS:2;f28094b72178:37967 {}] regionserver.Replication(140): Replication stats-in-log period=300 seconds 2025-01-08T04:58:11,211 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T04:58:11,212 INFO [PEWorker-1 {}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=60249958, jitterRate=-0.10220566391944885}}}, FlushLargeStoresPolicy{flushSizeLowerBound=44739242} 2025-01-08T04:58:11,215 DEBUG [PEWorker-1 {}] regionserver.HRegion(1001): Region open journal for 1588230740: 2025-01-08T04:58:11,215 DEBUG [PEWorker-1 {}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2025-01-08T04:58:11,215 INFO [PEWorker-1 {}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2025-01-08T04:58:11,215 DEBUG [PEWorker-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2025-01-08T04:58:11,215 DEBUG [PEWorker-1 {}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2025-01-08T04:58:11,215 DEBUG [PEWorker-1 {}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2025-01-08T04:58:11,216 INFO [PEWorker-1 {}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2025-01-08T04:58:11,217 DEBUG [PEWorker-1 {}] regionserver.HRegion(1635): Region close journal for 1588230740: 2025-01-08T04:58:11,220 INFO [RS:0;f28094b72178:39801 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2025-01-08T04:58:11,221 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, locked=true; InitMetaProcedure table=hbase:meta 2025-01-08T04:58:11,222 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(107): Going to assign meta 2025-01-08T04:58:11,222 INFO [RS:1;f28094b72178:37583 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2025-01-08T04:58:11,222 INFO [RS:2;f28094b72178:37967 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2025-01-08T04:58:11,230 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2025-01-08T04:58:11,237 INFO [RS:0;f28094b72178:39801 {}] regionserver.MemStoreFlusher(130): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2025-01-08T04:58:11,243 INFO [RS:2;f28094b72178:37967 {}] regionserver.MemStoreFlusher(130): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2025-01-08T04:58:11,249 INFO [RS:0;f28094b72178:39801 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2025-01-08T04:58:11,249 INFO [RS:0;f28094b72178:39801 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:11,263 INFO [RS:2;f28094b72178:37967 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2025-01-08T04:58:11,263 INFO [RS:2;f28094b72178:37967 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:11,273 INFO [RS:0;f28094b72178:39801 {}] regionserver.HRegionServer$CompactionChecker(1988): CompactionChecker runs every PT1S 2025-01-08T04:58:11,276 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2025-01-08T04:58:11,278 INFO [RS:1;f28094b72178:37583 {}] regionserver.MemStoreFlusher(130): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2025-01-08T04:58:11,281 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2025-01-08T04:58:11,289 INFO [RS:0;f28094b72178:39801 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:11,291 DEBUG [RS:0;f28094b72178:39801 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,292 INFO [RS:1;f28094b72178:37583 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2025-01-08T04:58:11,292 DEBUG [RS:0;f28094b72178:39801 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,292 INFO [RS:1;f28094b72178:37583 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:11,292 DEBUG [RS:0;f28094b72178:39801 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,292 DEBUG [RS:0;f28094b72178:39801 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,292 DEBUG [RS:0;f28094b72178:39801 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,293 DEBUG [RS:0;f28094b72178:39801 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/f28094b72178:0, corePoolSize=2, maxPoolSize=2 2025-01-08T04:58:11,293 DEBUG [RS:0;f28094b72178:39801 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,293 DEBUG [RS:0;f28094b72178:39801 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,293 DEBUG [RS:0;f28094b72178:39801 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,294 DEBUG [RS:0;f28094b72178:39801 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,295 DEBUG [RS:0;f28094b72178:39801 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,295 DEBUG [RS:0;f28094b72178:39801 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0, corePoolSize=3, maxPoolSize=3 2025-01-08T04:58:11,295 DEBUG [RS:0;f28094b72178:39801 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/f28094b72178:0, corePoolSize=3, maxPoolSize=3 2025-01-08T04:58:11,297 INFO [RS:1;f28094b72178:37583 {}] regionserver.HRegionServer$CompactionChecker(1988): CompactionChecker runs every PT1S 2025-01-08T04:58:11,300 INFO [RS:1;f28094b72178:37583 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:11,300 DEBUG [RS:1;f28094b72178:37583 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,300 DEBUG [RS:1;f28094b72178:37583 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,300 DEBUG [RS:1;f28094b72178:37583 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,301 DEBUG [RS:1;f28094b72178:37583 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,301 DEBUG [RS:1;f28094b72178:37583 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,301 DEBUG [RS:1;f28094b72178:37583 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/f28094b72178:0, corePoolSize=2, maxPoolSize=2 2025-01-08T04:58:11,302 DEBUG [RS:1;f28094b72178:37583 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,302 DEBUG [RS:1;f28094b72178:37583 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,302 DEBUG [RS:1;f28094b72178:37583 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,302 INFO [RS:2;f28094b72178:37967 {}] regionserver.HRegionServer$CompactionChecker(1988): CompactionChecker runs every PT1S 2025-01-08T04:58:11,302 DEBUG [RS:1;f28094b72178:37583 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,305 DEBUG [RS:1;f28094b72178:37583 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,305 DEBUG [RS:1;f28094b72178:37583 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0, corePoolSize=3, maxPoolSize=3 2025-01-08T04:58:11,306 DEBUG [RS:1;f28094b72178:37583 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/f28094b72178:0, corePoolSize=3, maxPoolSize=3 2025-01-08T04:58:11,306 INFO [RS:2;f28094b72178:37967 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:11,306 DEBUG [RS:2;f28094b72178:37967 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,307 DEBUG [RS:2;f28094b72178:37967 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,307 DEBUG [RS:2;f28094b72178:37967 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,307 DEBUG [RS:2;f28094b72178:37967 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,307 DEBUG [RS:2;f28094b72178:37967 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,307 DEBUG [RS:2;f28094b72178:37967 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/f28094b72178:0, corePoolSize=2, maxPoolSize=2 2025-01-08T04:58:11,307 DEBUG [RS:2;f28094b72178:37967 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,308 DEBUG [RS:2;f28094b72178:37967 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,308 DEBUG [RS:2;f28094b72178:37967 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,308 DEBUG [RS:2;f28094b72178:37967 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,308 DEBUG [RS:2;f28094b72178:37967 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/f28094b72178:0, corePoolSize=1, maxPoolSize=1 2025-01-08T04:58:11,308 DEBUG [RS:2;f28094b72178:37967 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0, corePoolSize=3, maxPoolSize=3 2025-01-08T04:58:11,308 DEBUG [RS:2;f28094b72178:37967 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/f28094b72178:0, corePoolSize=3, maxPoolSize=3 2025-01-08T04:58:11,322 INFO [RS:1;f28094b72178:37583 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:11,323 INFO [RS:1;f28094b72178:37583 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:11,323 INFO [RS:1;f28094b72178:37583 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:11,323 INFO [RS:1;f28094b72178:37583 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:11,323 INFO [RS:1;f28094b72178:37583 {}] hbase.ChoreService(168): Chore ScheduledChore name=f28094b72178,37583,1736312288873-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2025-01-08T04:58:11,361 INFO [RS:2;f28094b72178:37967 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:11,361 INFO [RS:2;f28094b72178:37967 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:11,361 INFO [RS:2;f28094b72178:37967 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:11,361 INFO [RS:2;f28094b72178:37967 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:11,361 INFO [RS:2;f28094b72178:37967 {}] hbase.ChoreService(168): Chore ScheduledChore name=f28094b72178,37967,1736312288941-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2025-01-08T04:58:11,367 INFO [RS:0;f28094b72178:39801 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:11,367 INFO [RS:0;f28094b72178:39801 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:11,367 INFO [RS:0;f28094b72178:39801 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:11,367 INFO [RS:0;f28094b72178:39801 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:11,367 INFO [RS:0;f28094b72178:39801 {}] hbase.ChoreService(168): Chore ScheduledChore name=f28094b72178,39801,1736312288779-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2025-01-08T04:58:11,395 INFO [RS:1;f28094b72178:37583 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2025-01-08T04:58:11,399 INFO [RS:1;f28094b72178:37583 {}] hbase.ChoreService(168): Chore ScheduledChore name=f28094b72178,37583,1736312288873-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:11,400 INFO [RS:2;f28094b72178:37967 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2025-01-08T04:58:11,401 INFO [RS:0;f28094b72178:39801 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2025-01-08T04:58:11,401 INFO [RS:2;f28094b72178:37967 {}] hbase.ChoreService(168): Chore ScheduledChore name=f28094b72178,37967,1736312288941-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:11,401 INFO [RS:0;f28094b72178:39801 {}] hbase.ChoreService(168): Chore ScheduledChore name=f28094b72178,39801,1736312288779-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:11,433 WARN [f28094b72178:42195 {}] assignment.AssignmentManager(2423): No servers available; cannot place 1 unassigned regions. 2025-01-08T04:58:11,434 INFO [RS:0;f28094b72178:39801 {}] regionserver.Replication(204): f28094b72178,39801,1736312288779 started 2025-01-08T04:58:11,435 INFO [RS:0;f28094b72178:39801 {}] regionserver.HRegionServer(1767): Serving as f28094b72178,39801,1736312288779, RpcServer on f28094b72178/172.17.0.3:39801, sessionid=0x1000ae1399f0001 2025-01-08T04:58:11,435 DEBUG [RS:0;f28094b72178:39801 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2025-01-08T04:58:11,436 DEBUG [RS:0;f28094b72178:39801 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager f28094b72178,39801,1736312288779 2025-01-08T04:58:11,436 DEBUG [RS:0;f28094b72178:39801 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'f28094b72178,39801,1736312288779' 2025-01-08T04:58:11,436 DEBUG [RS:0;f28094b72178:39801 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2025-01-08T04:58:11,437 DEBUG [RS:0;f28094b72178:39801 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2025-01-08T04:58:11,437 DEBUG [RS:0;f28094b72178:39801 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2025-01-08T04:58:11,437 DEBUG [RS:0;f28094b72178:39801 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2025-01-08T04:58:11,437 DEBUG [RS:0;f28094b72178:39801 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager f28094b72178,39801,1736312288779 2025-01-08T04:58:11,437 DEBUG [RS:0;f28094b72178:39801 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'f28094b72178,39801,1736312288779' 2025-01-08T04:58:11,437 DEBUG [RS:0;f28094b72178:39801 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2025-01-08T04:58:11,438 DEBUG [RS:0;f28094b72178:39801 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2025-01-08T04:58:11,438 INFO [RS:2;f28094b72178:37967 {}] regionserver.Replication(204): f28094b72178,37967,1736312288941 started 2025-01-08T04:58:11,438 INFO [RS:2;f28094b72178:37967 {}] regionserver.HRegionServer(1767): Serving as f28094b72178,37967,1736312288941, RpcServer on f28094b72178/172.17.0.3:37967, sessionid=0x1000ae1399f0003 2025-01-08T04:58:11,439 DEBUG [RS:2;f28094b72178:37967 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2025-01-08T04:58:11,439 DEBUG [RS:0;f28094b72178:39801 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2025-01-08T04:58:11,439 DEBUG [RS:2;f28094b72178:37967 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager f28094b72178,37967,1736312288941 2025-01-08T04:58:11,439 DEBUG [RS:2;f28094b72178:37967 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'f28094b72178,37967,1736312288941' 2025-01-08T04:58:11,439 INFO [RS:0;f28094b72178:39801 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2025-01-08T04:58:11,439 DEBUG [RS:2;f28094b72178:37967 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2025-01-08T04:58:11,439 INFO [RS:0;f28094b72178:39801 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2025-01-08T04:58:11,439 DEBUG [RS:2;f28094b72178:37967 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2025-01-08T04:58:11,440 DEBUG [RS:2;f28094b72178:37967 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2025-01-08T04:58:11,440 DEBUG [RS:2;f28094b72178:37967 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2025-01-08T04:58:11,440 DEBUG [RS:2;f28094b72178:37967 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager f28094b72178,37967,1736312288941 2025-01-08T04:58:11,440 DEBUG [RS:2;f28094b72178:37967 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'f28094b72178,37967,1736312288941' 2025-01-08T04:58:11,440 DEBUG [RS:2;f28094b72178:37967 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2025-01-08T04:58:11,444 INFO [RS:1;f28094b72178:37583 {}] regionserver.Replication(204): f28094b72178,37583,1736312288873 started 2025-01-08T04:58:11,444 INFO [RS:1;f28094b72178:37583 {}] regionserver.HRegionServer(1767): Serving as f28094b72178,37583,1736312288873, RpcServer on f28094b72178/172.17.0.3:37583, sessionid=0x1000ae1399f0002 2025-01-08T04:58:11,444 DEBUG [RS:1;f28094b72178:37583 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2025-01-08T04:58:11,444 DEBUG [RS:1;f28094b72178:37583 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager f28094b72178,37583,1736312288873 2025-01-08T04:58:11,444 DEBUG [RS:1;f28094b72178:37583 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'f28094b72178,37583,1736312288873' 2025-01-08T04:58:11,444 DEBUG [RS:1;f28094b72178:37583 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2025-01-08T04:58:11,459 DEBUG [RS:1;f28094b72178:37583 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2025-01-08T04:58:11,459 DEBUG [RS:2;f28094b72178:37967 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2025-01-08T04:58:11,467 DEBUG [RS:1;f28094b72178:37583 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2025-01-08T04:58:11,467 DEBUG [RS:1;f28094b72178:37583 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2025-01-08T04:58:11,467 DEBUG [RS:1;f28094b72178:37583 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager f28094b72178,37583,1736312288873 2025-01-08T04:58:11,467 DEBUG [RS:1;f28094b72178:37583 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'f28094b72178,37583,1736312288873' 2025-01-08T04:58:11,467 DEBUG [RS:1;f28094b72178:37583 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2025-01-08T04:58:11,467 DEBUG [RS:2;f28094b72178:37967 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2025-01-08T04:58:11,467 INFO [RS:2;f28094b72178:37967 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2025-01-08T04:58:11,467 INFO [RS:2;f28094b72178:37967 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2025-01-08T04:58:11,468 DEBUG [RS:1;f28094b72178:37583 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2025-01-08T04:58:11,469 DEBUG [RS:1;f28094b72178:37583 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2025-01-08T04:58:11,469 INFO [RS:1;f28094b72178:37583 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2025-01-08T04:58:11,469 INFO [RS:1;f28094b72178:37583 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2025-01-08T04:58:11,546 INFO [RS:0;f28094b72178:39801 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2025-01-08T04:58:11,552 INFO [RS:0;f28094b72178:39801 {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=f28094b72178%2C39801%2C1736312288779, suffix=, logDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/WALs/f28094b72178,39801,1736312288779, archiveDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/oldWALs, maxLogs=32 2025-01-08T04:58:11,568 INFO [RS:2;f28094b72178:37967 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2025-01-08T04:58:11,569 INFO [RS:1;f28094b72178:37583 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2025-01-08T04:58:11,572 INFO [RS:2;f28094b72178:37967 {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=f28094b72178%2C37967%2C1736312288941, suffix=, logDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/WALs/f28094b72178,37967,1736312288941, archiveDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/oldWALs, maxLogs=32 2025-01-08T04:58:11,572 INFO [RS:1;f28094b72178:37583 {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=f28094b72178%2C37583%2C1736312288873, suffix=, logDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/WALs/f28094b72178,37583,1736312288873, archiveDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/oldWALs, maxLogs=32 2025-01-08T04:58:11,576 DEBUG [RS:0;f28094b72178:39801 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(617): When create output stream for /user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/WALs/f28094b72178,39801,1736312288779/f28094b72178%2C39801%2C1736312288779.1736312291555, exclude list is [], retry=0 2025-01-08T04:58:11,581 DEBUG [RS-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:40353,DS-0a2953dd-a4c6-4650-9631-02981d9d1e8d,DISK] 2025-01-08T04:58:11,582 DEBUG [RS-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:33991,DS-f676faff-d67d-404f-bb0f-b90001fc3639,DISK] 2025-01-08T04:58:11,582 DEBUG [RS-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45867,DS-84b09c6f-eff8-44d2-9077-117e68175eb1,DISK] 2025-01-08T04:58:11,592 INFO [RS:0;f28094b72178:39801 {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/WALs/f28094b72178,39801,1736312288779/f28094b72178%2C39801%2C1736312288779.1736312291555 2025-01-08T04:58:11,593 DEBUG [RS:0;f28094b72178:39801 {}] wal.AbstractFSWAL(925): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:33789:33789),(127.0.0.1/127.0.0.1:45387:45387),(127.0.0.1/127.0.0.1:34673:34673)] 2025-01-08T04:58:11,595 DEBUG [RS:2;f28094b72178:37967 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(617): When create output stream for /user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/WALs/f28094b72178,37967,1736312288941/f28094b72178%2C37967%2C1736312288941.1736312291574, exclude list is [], retry=0 2025-01-08T04:58:11,596 DEBUG [RS:1;f28094b72178:37583 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(617): When create output stream for /user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/WALs/f28094b72178,37583,1736312288873/f28094b72178%2C37583%2C1736312288873.1736312291574, exclude list is [], retry=0 2025-01-08T04:58:11,603 DEBUG [RS-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:40353,DS-0a2953dd-a4c6-4650-9631-02981d9d1e8d,DISK] 2025-01-08T04:58:11,603 DEBUG [RS-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45867,DS-84b09c6f-eff8-44d2-9077-117e68175eb1,DISK] 2025-01-08T04:58:11,604 DEBUG [RS-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:33991,DS-f676faff-d67d-404f-bb0f-b90001fc3639,DISK] 2025-01-08T04:58:11,607 DEBUG [RS-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45867,DS-84b09c6f-eff8-44d2-9077-117e68175eb1,DISK] 2025-01-08T04:58:11,607 DEBUG [RS-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:40353,DS-0a2953dd-a4c6-4650-9631-02981d9d1e8d,DISK] 2025-01-08T04:58:11,607 DEBUG [RS-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:33991,DS-f676faff-d67d-404f-bb0f-b90001fc3639,DISK] 2025-01-08T04:58:11,628 INFO [RS:1;f28094b72178:37583 {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/WALs/f28094b72178,37583,1736312288873/f28094b72178%2C37583%2C1736312288873.1736312291574 2025-01-08T04:58:11,628 INFO [RS:2;f28094b72178:37967 {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/WALs/f28094b72178,37967,1736312288941/f28094b72178%2C37967%2C1736312288941.1736312291574 2025-01-08T04:58:11,630 DEBUG [RS:1;f28094b72178:37583 {}] wal.AbstractFSWAL(925): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:34673:34673),(127.0.0.1/127.0.0.1:33789:33789),(127.0.0.1/127.0.0.1:45387:45387)] 2025-01-08T04:58:11,634 DEBUG [RS:2;f28094b72178:37967 {}] wal.AbstractFSWAL(925): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:33789:33789),(127.0.0.1/127.0.0.1:34673:34673),(127.0.0.1/127.0.0.1:45387:45387)] 2025-01-08T04:58:11,685 DEBUG [f28094b72178:42195 {}] assignment.AssignmentManager(2444): Processing assignQueue; systemServersCount=3, allServersCount=3 2025-01-08T04:58:11,688 DEBUG [f28094b72178:42195 {}] balancer.BalancerClusterState(202): Hosts are {f28094b72178=0} racks are {/default-rack=0} 2025-01-08T04:58:11,697 DEBUG [f28094b72178:42195 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T04:58:11,697 DEBUG [f28094b72178:42195 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T04:58:11,698 DEBUG [f28094b72178:42195 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T04:58:11,698 INFO [f28094b72178:42195 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T04:58:11,698 INFO [f28094b72178:42195 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T04:58:11,698 INFO [f28094b72178:42195 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T04:58:11,698 DEBUG [f28094b72178:42195 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T04:58:11,712 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T04:58:11,719 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as f28094b72178,37583,1736312288873, state=OPENING 2025-01-08T04:58:11,726 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2025-01-08T04:58:11,728 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:11,729 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:11,729 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:11,730 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T04:58:11,730 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T04:58:11,730 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:11,731 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T04:58:11,731 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T04:58:11,735 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE; OpenRegionProcedure 1588230740, server=f28094b72178,37583,1736312288873}] 2025-01-08T04:58:11,923 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T04:58:11,925 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2025-01-08T04:58:11,931 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:55242, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2025-01-08T04:58:11,968 INFO [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(135): Open hbase:meta,,1.1588230740 2025-01-08T04:58:11,968 INFO [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2025-01-08T04:58:11,969 INFO [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor .meta 2025-01-08T04:58:11,974 INFO [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=f28094b72178%2C37583%2C1736312288873.meta, suffix=.meta, logDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/WALs/f28094b72178,37583,1736312288873, archiveDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/oldWALs, maxLogs=32 2025-01-08T04:58:12,001 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=3}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(617): When create output stream for /user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/WALs/f28094b72178,37583,1736312288873/f28094b72178%2C37583%2C1736312288873.meta.1736312291976.meta, exclude list is [], retry=0 2025-01-08T04:58:12,012 DEBUG [RS-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:40353,DS-0a2953dd-a4c6-4650-9631-02981d9d1e8d,DISK] 2025-01-08T04:58:12,014 DEBUG [RS-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:33991,DS-f676faff-d67d-404f-bb0f-b90001fc3639,DISK] 2025-01-08T04:58:12,016 DEBUG [RS-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45867,DS-84b09c6f-eff8-44d2-9077-117e68175eb1,DISK] 2025-01-08T04:58:12,027 INFO [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/WALs/f28094b72178,37583,1736312288873/f28094b72178%2C37583%2C1736312288873.meta.1736312291976.meta 2025-01-08T04:58:12,030 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(925): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:45387:45387),(127.0.0.1/127.0.0.1:33789:33789),(127.0.0.1/127.0.0.1:34673:34673)] 2025-01-08T04:58:12,031 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7285): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2025-01-08T04:58:12,033 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7999): Registered coprocessor service: region=hbase:meta,,1 service=AccessControlService 2025-01-08T04:58:12,034 INFO [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T04:58:12,036 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2025-01-08T04:58:12,040 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7999): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2025-01-08T04:58:12,042 INFO [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(436): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2025-01-08T04:58:12,060 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2025-01-08T04:58:12,060 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T04:58:12,061 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7327): checking encryption for 1588230740 2025-01-08T04:58:12,061 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7330): checking classloading for 1588230740 2025-01-08T04:58:12,066 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2025-01-08T04:58:12,069 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2025-01-08T04:58:12,069 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:58:12,071 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2025-01-08T04:58:12,071 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2025-01-08T04:58:12,072 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2025-01-08T04:58:12,073 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:58:12,074 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2025-01-08T04:58:12,074 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2025-01-08T04:58:12,078 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2025-01-08T04:58:12,079 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:58:12,080 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2025-01-08T04:58:12,082 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/1588230740 2025-01-08T04:58:12,085 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/1588230740 2025-01-08T04:58:12,088 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (42.7 M)) instead. 2025-01-08T04:58:12,092 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1085): writing seq id for 1588230740 2025-01-08T04:58:12,095 INFO [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=70796692, jitterRate=0.054952919483184814}}}, FlushLargeStoresPolicy{flushSizeLowerBound=44739242} 2025-01-08T04:58:12,099 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1001): Region open journal for 1588230740: 2025-01-08T04:58:12,108 INFO [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1736312291913 2025-01-08T04:58:12,123 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:meta,,1.1588230740 2025-01-08T04:58:12,123 INFO [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(164): Opened hbase:meta,,1.1588230740 2025-01-08T04:58:12,125 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T04:58:12,128 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as f28094b72178,37583,1736312288873, state=OPEN 2025-01-08T04:58:12,135 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2025-01-08T04:58:12,135 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2025-01-08T04:58:12,135 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2025-01-08T04:58:12,135 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T04:58:12,135 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T04:58:12,135 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T04:58:12,140 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2025-01-08T04:58:12,140 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T04:58:12,146 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=3, resume processing ppid=2 2025-01-08T04:58:12,147 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=3, ppid=2, state=SUCCESS; OpenRegionProcedure 1588230740, server=f28094b72178,37583,1736312288873 in 400 msec 2025-01-08T04:58:12,155 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=2, resume processing ppid=1 2025-01-08T04:58:12,155 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=2, ppid=1, state=SUCCESS; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 918 msec 2025-01-08T04:58:12,165 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=1, state=SUCCESS; InitMetaProcedure table=hbase:meta in 1.3100 sec 2025-01-08T04:58:12,165 INFO [master/f28094b72178:0:becomeActiveMaster {}] master.HMaster(1099): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1736312292165, completionTime=-1 2025-01-08T04:58:12,165 INFO [master/f28094b72178:0:becomeActiveMaster {}] master.ServerManager(907): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2025-01-08T04:58:12,165 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] assignment.AssignmentManager(1747): Joining cluster... 2025-01-08T04:58:12,209 DEBUG [hconnection-0x168697cd-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T04:58:12,225 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:55258, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T04:58:12,255 INFO [master/f28094b72178:0:becomeActiveMaster {}] assignment.AssignmentManager(1759): Number of RegionServers=3 2025-01-08T04:58:12,256 INFO [master/f28094b72178:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1736312352255 2025-01-08T04:58:12,256 INFO [master/f28094b72178:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1736312412256 2025-01-08T04:58:12,256 INFO [master/f28094b72178:0:becomeActiveMaster {}] assignment.AssignmentManager(1766): Joined the cluster in 90 msec 2025-01-08T04:58:12,288 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] balancer.RegionLocationFinder(172): Locality for region 1588230740 changed from -1.0 to 0.0, refreshing cache 2025-01-08T04:58:12,312 INFO [master/f28094b72178:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=f28094b72178,42195,1736312287279-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:12,312 INFO [master/f28094b72178:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=f28094b72178,42195,1736312287279-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:12,313 INFO [master/f28094b72178:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=f28094b72178,42195,1736312287279-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:12,315 INFO [master/f28094b72178:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-f28094b72178:42195, period=300000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:12,315 INFO [master/f28094b72178:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:12,323 DEBUG [master/f28094b72178:0.Chore.1 {}] janitor.CatalogJanitor(179): 2025-01-08T04:58:12,326 INFO [master/f28094b72178:0:becomeActiveMaster {}] master.TableNamespaceManager(92): Namespace table not found. Creating... 2025-01-08T04:58:12,327 INFO [master/f28094b72178:0:becomeActiveMaster {}] master.HMaster(2440): Client=null/null create 'hbase:namespace', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2025-01-08T04:58:12,334 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=hbase:namespace 2025-01-08T04:58:12,340 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T04:58:12,343 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:58:12,346 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T04:58:12,378 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741837_1013 (size=358) 2025-01-08T04:58:12,378 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741837_1013 (size=358) 2025-01-08T04:58:12,379 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741837_1013 (size=358) 2025-01-08T04:58:12,793 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 52312176a1351392d55e81604512c13b, NAME => 'hbase:namespace,,1736312292327.52312176a1351392d55e81604512c13b.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:namespace', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T04:58:12,841 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741838_1014 (size=42) 2025-01-08T04:58:12,842 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741838_1014 (size=42) 2025-01-08T04:58:12,842 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741838_1014 (size=42) 2025-01-08T04:58:12,844 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(894): Instantiated hbase:namespace,,1736312292327.52312176a1351392d55e81604512c13b.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T04:58:12,844 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1681): Closing 52312176a1351392d55e81604512c13b, disabling compactions & flushes 2025-01-08T04:58:12,844 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1703): Closing region hbase:namespace,,1736312292327.52312176a1351392d55e81604512c13b. 2025-01-08T04:58:12,844 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1736312292327.52312176a1351392d55e81604512c13b. 2025-01-08T04:58:12,844 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1736312292327.52312176a1351392d55e81604512c13b. after waiting 0 ms 2025-01-08T04:58:12,844 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1736312292327.52312176a1351392d55e81604512c13b. 2025-01-08T04:58:12,844 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1922): Closed hbase:namespace,,1736312292327.52312176a1351392d55e81604512c13b. 2025-01-08T04:58:12,844 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1635): Region close journal for 52312176a1351392d55e81604512c13b: 2025-01-08T04:58:12,847 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T04:58:12,854 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"hbase:namespace,,1736312292327.52312176a1351392d55e81604512c13b.","families":{"info":[{"qualifier":"regioninfo","vlen":41,"tag":[],"timestamp":"1736312292848"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736312292848"}]},"ts":"1736312292848"} 2025-01-08T04:58:12,886 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2025-01-08T04:58:12,889 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T04:58:12,892 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312292889"}]},"ts":"1736312292889"} 2025-01-08T04:58:12,898 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLING in hbase:meta 2025-01-08T04:58:12,903 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(202): Hosts are {f28094b72178=0} racks are {/default-rack=0} 2025-01-08T04:58:12,905 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T04:58:12,905 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T04:58:12,905 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T04:58:12,905 INFO [PEWorker-3 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T04:58:12,905 INFO [PEWorker-3 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T04:58:12,905 INFO [PEWorker-3 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T04:58:12,905 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T04:58:12,907 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=52312176a1351392d55e81604512c13b, ASSIGN}] 2025-01-08T04:58:12,910 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=52312176a1351392d55e81604512c13b, ASSIGN 2025-01-08T04:58:12,913 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:namespace, region=52312176a1351392d55e81604512c13b, ASSIGN; state=OFFLINE, location=f28094b72178,37583,1736312288873; forceNewPlan=false, retain=false 2025-01-08T04:58:13,065 INFO [f28094b72178:42195 {}] balancer.BaseLoadBalancer(546): Reassigned 1 regions. 1 retained the pre-restart assignment. 2025-01-08T04:58:13,065 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=52312176a1351392d55e81604512c13b, regionState=OPENING, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T04:58:13,083 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE; OpenRegionProcedure 52312176a1351392d55e81604512c13b, server=f28094b72178,37583,1736312288873}] 2025-01-08T04:58:13,245 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T04:58:13,252 INFO [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(135): Open hbase:namespace,,1736312292327.52312176a1351392d55e81604512c13b. 2025-01-08T04:58:13,253 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7285): Opening region: {ENCODED => 52312176a1351392d55e81604512c13b, NAME => 'hbase:namespace,,1736312292327.52312176a1351392d55e81604512c13b.', STARTKEY => '', ENDKEY => ''} 2025-01-08T04:58:13,253 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7999): Registered coprocessor service: region=hbase:namespace,,1736312292327.52312176a1351392d55e81604512c13b. service=AccessControlService 2025-01-08T04:58:13,253 INFO [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T04:58:13,254 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table namespace 52312176a1351392d55e81604512c13b 2025-01-08T04:58:13,254 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(894): Instantiated hbase:namespace,,1736312292327.52312176a1351392d55e81604512c13b.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T04:58:13,254 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7327): checking encryption for 52312176a1351392d55e81604512c13b 2025-01-08T04:58:13,254 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7330): checking classloading for 52312176a1351392d55e81604512c13b 2025-01-08T04:58:13,257 INFO [StoreOpener-52312176a1351392d55e81604512c13b-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 52312176a1351392d55e81604512c13b 2025-01-08T04:58:13,260 INFO [StoreOpener-52312176a1351392d55e81604512c13b-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 52312176a1351392d55e81604512c13b columnFamilyName info 2025-01-08T04:58:13,261 DEBUG [StoreOpener-52312176a1351392d55e81604512c13b-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:58:13,262 INFO [StoreOpener-52312176a1351392d55e81604512c13b-1 {}] regionserver.HStore(327): Store=52312176a1351392d55e81604512c13b/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T04:58:13,264 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/namespace/52312176a1351392d55e81604512c13b 2025-01-08T04:58:13,264 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/namespace/52312176a1351392d55e81604512c13b 2025-01-08T04:58:13,269 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1085): writing seq id for 52312176a1351392d55e81604512c13b 2025-01-08T04:58:13,273 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/namespace/52312176a1351392d55e81604512c13b/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T04:58:13,274 INFO [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1102): Opened 52312176a1351392d55e81604512c13b; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=73843315, jitterRate=0.10035113990306854}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T04:58:13,277 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1001): Region open journal for 52312176a1351392d55e81604512c13b: 2025-01-08T04:58:13,283 INFO [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:namespace,,1736312292327.52312176a1351392d55e81604512c13b., pid=6, masterSystemTime=1736312293245 2025-01-08T04:58:13,289 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:namespace,,1736312292327.52312176a1351392d55e81604512c13b. 2025-01-08T04:58:13,289 INFO [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(164): Opened hbase:namespace,,1736312292327.52312176a1351392d55e81604512c13b. 2025-01-08T04:58:13,290 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=52312176a1351392d55e81604512c13b, regionState=OPEN, openSeqNum=2, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T04:58:13,298 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=6, resume processing ppid=5 2025-01-08T04:58:13,299 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=6, ppid=5, state=SUCCESS; OpenRegionProcedure 52312176a1351392d55e81604512c13b, server=f28094b72178,37583,1736312288873 in 211 msec 2025-01-08T04:58:13,302 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=5, resume processing ppid=4 2025-01-08T04:58:13,302 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=5, ppid=4, state=SUCCESS; TransitRegionStateProcedure table=hbase:namespace, region=52312176a1351392d55e81604512c13b, ASSIGN in 391 msec 2025-01-08T04:58:13,303 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T04:58:13,304 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312293304"}]},"ts":"1736312293304"} 2025-01-08T04:58:13,307 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLED in hbase:meta 2025-01-08T04:58:13,311 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T04:58:13,315 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=4, state=SUCCESS; CreateTableProcedure table=hbase:namespace in 983 msec 2025-01-08T04:58:13,341 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] zookeeper.ZKUtil(113): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/namespace 2025-01-08T04:58:13,343 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/namespace 2025-01-08T04:58:13,343 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:13,343 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:13,343 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:13,343 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:13,471 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=7, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=default 2025-01-08T04:58:13,532 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2025-01-08T04:58:13,550 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=7, state=SUCCESS; CreateNamespaceProcedure, namespace=default in 86 msec 2025-01-08T04:58:13,558 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=8, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=hbase 2025-01-08T04:58:13,578 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2025-01-08T04:58:13,586 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=8, state=SUCCESS; CreateNamespaceProcedure, namespace=hbase in 29 msec 2025-01-08T04:58:13,598 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/default 2025-01-08T04:58:13,602 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/hbase 2025-01-08T04:58:13,602 INFO [master/f28094b72178:0:becomeActiveMaster {}] master.HMaster(1229): Master has completed initialization 4.575sec 2025-01-08T04:58:13,620 INFO [master/f28094b72178:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2025-01-08T04:58:13,622 INFO [master/f28094b72178:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2025-01-08T04:58:13,624 INFO [master/f28094b72178:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2025-01-08T04:58:13,625 INFO [master/f28094b72178:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2025-01-08T04:58:13,625 INFO [master/f28094b72178:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2025-01-08T04:58:13,626 INFO [master/f28094b72178:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=f28094b72178,42195,1736312287279-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2025-01-08T04:58:13,627 INFO [master/f28094b72178:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=f28094b72178,42195,1736312287279-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2025-01-08T04:58:13,659 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(149): Connect 0x4476f33e to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@654ff23 2025-01-08T04:58:13,666 INFO [master/f28094b72178:0:becomeActiveMaster {}] master.HMaster$4(2404): Client=null/null create 'hbase:acl', {NAME => 'l', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2025-01-08T04:58:13,674 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=9, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=hbase:acl 2025-01-08T04:58:13,678 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=hbase:acl execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T04:58:13,679 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:58:13,679 INFO [master/f28094b72178:0:becomeActiveMaster {}] master.MasterRpcServices(713): Client=null/null procedure request for creating table: namespace: "hbase" qualifier: "acl" procId is: 9 2025-01-08T04:58:13,681 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=hbase:acl execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T04:58:13,689 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2025-01-08T04:58:13,690 WARN [Time-limited test {}] client.ZKConnectionRegistry(90): ZKConnectionRegistry is deprecated. See https://hbase.apache.org/book.html#client.rpcconnectionregistry 2025-01-08T04:58:13,732 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@70688cdf, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T04:58:13,759 DEBUG [Time-limited test {}] nio.NioEventLoop(110): -Dio.netty.noKeySetOptimization: false 2025-01-08T04:58:13,759 DEBUG [Time-limited test {}] nio.NioEventLoop(111): -Dio.netty.selectorAutoRebuildThreshold: 512 2025-01-08T04:58:13,779 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741839_1015 (size=349) 2025-01-08T04:58:13,780 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741839_1015 (size=349) 2025-01-08T04:58:13,781 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741839_1015 (size=349) 2025-01-08T04:58:13,789 DEBUG [hconnection-0x39b30dcc-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T04:58:13,790 INFO [RegionOpenAndInit-hbase:acl-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => c1df98ea008addbfc51d7e8682ea8fd9, NAME => 'hbase:acl,,1736312293651.c1df98ea008addbfc51d7e8682ea8fd9.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:acl', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'l', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T04:58:13,799 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2025-01-08T04:58:13,817 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:55274, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T04:58:13,821 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1199): Minicluster is up; activeMaster=f28094b72178,42195,1736312287279 2025-01-08T04:58:13,821 INFO [Time-limited test {}] hbase.HBaseTestingUtility(2790): Starting mini mapreduce cluster... 2025-01-08T04:58:13,822 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/test.cache.data in system properties and HBase conf 2025-01-08T04:58:13,822 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop.tmp.dir in system properties and HBase conf 2025-01-08T04:58:13,822 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop.log.dir in system properties and HBase conf 2025-01-08T04:58:13,822 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/mapreduce.cluster.local.dir in system properties and HBase conf 2025-01-08T04:58:13,822 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/mapreduce.cluster.temp.dir in system properties and HBase conf 2025-01-08T04:58:13,822 INFO [Time-limited test {}] hbase.HBaseTestingUtility(811): read short circuit is OFF 2025-01-08T04:58:13,822 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2025-01-08T04:58:13,822 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2025-01-08T04:58:13,822 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/yarn.nodemanager.log-dirs in system properties and HBase conf 2025-01-08T04:58:13,822 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2025-01-08T04:58:13,823 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2025-01-08T04:58:13,823 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2025-01-08T04:58:13,823 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2025-01-08T04:58:13,823 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/dfs.journalnode.edits.dir in system properties and HBase conf 2025-01-08T04:58:13,823 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2025-01-08T04:58:13,823 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/nfs.dump.dir in system properties and HBase conf 2025-01-08T04:58:13,823 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/java.io.tmpdir in system properties and HBase conf 2025-01-08T04:58:13,823 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/dfs.journalnode.edits.dir in system properties and HBase conf 2025-01-08T04:58:13,823 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2025-01-08T04:58:13,823 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2025-01-08T04:58:13,830 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741840_1016 (size=36) 2025-01-08T04:58:13,836 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741840_1016 (size=36) 2025-01-08T04:58:13,843 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741840_1016 (size=36) 2025-01-08T04:58:13,845 DEBUG [RegionOpenAndInit-hbase:acl-pool-0 {}] regionserver.HRegion(894): Instantiated hbase:acl,,1736312293651.c1df98ea008addbfc51d7e8682ea8fd9.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T04:58:13,845 DEBUG [RegionOpenAndInit-hbase:acl-pool-0 {}] regionserver.HRegion(1681): Closing c1df98ea008addbfc51d7e8682ea8fd9, disabling compactions & flushes 2025-01-08T04:58:13,845 INFO [RegionOpenAndInit-hbase:acl-pool-0 {}] regionserver.HRegion(1703): Closing region hbase:acl,,1736312293651.c1df98ea008addbfc51d7e8682ea8fd9. 2025-01-08T04:58:13,845 DEBUG [RegionOpenAndInit-hbase:acl-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:acl,,1736312293651.c1df98ea008addbfc51d7e8682ea8fd9. 2025-01-08T04:58:13,845 DEBUG [RegionOpenAndInit-hbase:acl-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on hbase:acl,,1736312293651.c1df98ea008addbfc51d7e8682ea8fd9. after waiting 0 ms 2025-01-08T04:58:13,846 DEBUG [RegionOpenAndInit-hbase:acl-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region hbase:acl,,1736312293651.c1df98ea008addbfc51d7e8682ea8fd9. 2025-01-08T04:58:13,846 INFO [RegionOpenAndInit-hbase:acl-pool-0 {}] regionserver.HRegion(1922): Closed hbase:acl,,1736312293651.c1df98ea008addbfc51d7e8682ea8fd9. 2025-01-08T04:58:13,846 DEBUG [RegionOpenAndInit-hbase:acl-pool-0 {}] regionserver.HRegion(1635): Region close journal for c1df98ea008addbfc51d7e8682ea8fd9: 2025-01-08T04:58:13,848 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=hbase:acl execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T04:58:13,849 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"hbase:acl,,1736312293651.c1df98ea008addbfc51d7e8682ea8fd9.","families":{"info":[{"qualifier":"regioninfo","vlen":35,"tag":[],"timestamp":"1736312293848"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736312293848"}]},"ts":"1736312293848"} 2025-01-08T04:58:13,856 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2025-01-08T04:58:13,859 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=hbase:acl execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T04:58:13,859 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:acl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312293859"}]},"ts":"1736312293859"} 2025-01-08T04:58:13,863 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:acl, state=ENABLING in hbase:meta 2025-01-08T04:58:13,869 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(202): Hosts are {f28094b72178=0} racks are {/default-rack=0} 2025-01-08T04:58:13,872 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T04:58:13,872 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T04:58:13,872 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T04:58:13,872 INFO [PEWorker-1 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T04:58:13,872 INFO [PEWorker-1 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T04:58:13,872 INFO [PEWorker-1 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T04:58:13,872 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T04:58:13,873 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:acl, region=c1df98ea008addbfc51d7e8682ea8fd9, ASSIGN}] 2025-01-08T04:58:13,877 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:acl, region=c1df98ea008addbfc51d7e8682ea8fd9, ASSIGN 2025-01-08T04:58:13,880 INFO [PEWorker-3 {}] assignment.TransitRegionStateProcedure(264): Starting pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:acl, region=c1df98ea008addbfc51d7e8682ea8fd9, ASSIGN; state=OFFLINE, location=f28094b72178,37583,1736312288873; forceNewPlan=false, retain=false 2025-01-08T04:58:13,966 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741841_1017 (size=592039) 2025-01-08T04:58:13,967 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741841_1017 (size=592039) 2025-01-08T04:58:13,967 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741841_1017 (size=592039) 2025-01-08T04:58:14,000 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2025-01-08T04:58:14,030 INFO [f28094b72178:42195 {}] balancer.BaseLoadBalancer(546): Reassigned 1 regions. 1 retained the pre-restart assignment. 2025-01-08T04:58:14,031 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=10 updating hbase:meta row=c1df98ea008addbfc51d7e8682ea8fd9, regionState=OPENING, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T04:58:14,049 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=11, ppid=10, state=RUNNABLE; OpenRegionProcedure c1df98ea008addbfc51d7e8682ea8fd9, server=f28094b72178,37583,1736312288873}] 2025-01-08T04:58:14,072 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741842_1018 (size=1663647) 2025-01-08T04:58:14,073 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741842_1018 (size=1663647) 2025-01-08T04:58:14,073 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741842_1018 (size=1663647) 2025-01-08T04:58:14,237 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T04:58:14,296 INFO [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] handler.AssignRegionHandler(135): Open hbase:acl,,1736312293651.c1df98ea008addbfc51d7e8682ea8fd9. 2025-01-08T04:58:14,297 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(7285): Opening region: {ENCODED => c1df98ea008addbfc51d7e8682ea8fd9, NAME => 'hbase:acl,,1736312293651.c1df98ea008addbfc51d7e8682ea8fd9.', STARTKEY => '', ENDKEY => ''} 2025-01-08T04:58:14,298 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(7999): Registered coprocessor service: region=hbase:acl,,1736312293651.c1df98ea008addbfc51d7e8682ea8fd9. service=AccessControlService 2025-01-08T04:58:14,298 INFO [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T04:58:14,298 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table acl c1df98ea008addbfc51d7e8682ea8fd9 2025-01-08T04:58:14,298 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(894): Instantiated hbase:acl,,1736312293651.c1df98ea008addbfc51d7e8682ea8fd9.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T04:58:14,299 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(7327): checking encryption for c1df98ea008addbfc51d7e8682ea8fd9 2025-01-08T04:58:14,299 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(7330): checking classloading for c1df98ea008addbfc51d7e8682ea8fd9 2025-01-08T04:58:14,300 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2025-01-08T04:58:14,306 INFO [StoreOpener-c1df98ea008addbfc51d7e8682ea8fd9-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family l of region c1df98ea008addbfc51d7e8682ea8fd9 2025-01-08T04:58:14,315 INFO [StoreOpener-c1df98ea008addbfc51d7e8682ea8fd9-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region c1df98ea008addbfc51d7e8682ea8fd9 columnFamilyName l 2025-01-08T04:58:14,315 DEBUG [StoreOpener-c1df98ea008addbfc51d7e8682ea8fd9-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:58:14,317 INFO [StoreOpener-c1df98ea008addbfc51d7e8682ea8fd9-1 {}] regionserver.HStore(327): Store=c1df98ea008addbfc51d7e8682ea8fd9/l, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T04:58:14,320 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/acl/c1df98ea008addbfc51d7e8682ea8fd9 2025-01-08T04:58:14,321 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/acl/c1df98ea008addbfc51d7e8682ea8fd9 2025-01-08T04:58:14,327 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(1085): writing seq id for c1df98ea008addbfc51d7e8682ea8fd9 2025-01-08T04:58:14,339 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/acl/c1df98ea008addbfc51d7e8682ea8fd9/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T04:58:14,341 INFO [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(1102): Opened c1df98ea008addbfc51d7e8682ea8fd9; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=68983616, jitterRate=0.02793598175048828}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T04:58:14,347 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(1001): Region open journal for c1df98ea008addbfc51d7e8682ea8fd9: 2025-01-08T04:58:14,350 INFO [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:acl,,1736312293651.c1df98ea008addbfc51d7e8682ea8fd9., pid=11, masterSystemTime=1736312294237 2025-01-08T04:58:14,356 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:acl,,1736312293651.c1df98ea008addbfc51d7e8682ea8fd9. 2025-01-08T04:58:14,357 INFO [RS_OPEN_PRIORITY_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] handler.AssignRegionHandler(164): Opened hbase:acl,,1736312293651.c1df98ea008addbfc51d7e8682ea8fd9. 2025-01-08T04:58:14,359 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=10 updating hbase:meta row=c1df98ea008addbfc51d7e8682ea8fd9, regionState=OPEN, openSeqNum=2, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T04:58:14,369 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=11, resume processing ppid=10 2025-01-08T04:58:14,369 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=11, ppid=10, state=SUCCESS; OpenRegionProcedure c1df98ea008addbfc51d7e8682ea8fd9, server=f28094b72178,37583,1736312288873 in 315 msec 2025-01-08T04:58:14,374 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=10, resume processing ppid=9 2025-01-08T04:58:14,374 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=10, ppid=9, state=SUCCESS; TransitRegionStateProcedure table=hbase:acl, region=c1df98ea008addbfc51d7e8682ea8fd9, ASSIGN in 496 msec 2025-01-08T04:58:14,377 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=hbase:acl execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T04:58:14,378 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:acl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312294377"}]},"ts":"1736312294377"} 2025-01-08T04:58:14,381 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:acl, state=ENABLED in hbase:meta 2025-01-08T04:58:14,410 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=hbase:acl execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T04:58:14,416 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=9, state=SUCCESS; CreateTableProcedure table=hbase:acl in 744 msec 2025-01-08T04:58:14,800 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2025-01-08T04:58:14,801 INFO [master/f28094b72178:0:becomeActiveMaster {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: hbase:acl, procId: 9 completed 2025-01-08T04:58:14,805 DEBUG [master/f28094b72178:0:becomeActiveMaster {}] master.HMaster(1332): Balancer post startup initialization complete, took 0 seconds 2025-01-08T04:58:14,807 INFO [master/f28094b72178:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2025-01-08T04:58:14,807 INFO [master/f28094b72178:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=f28094b72178,42195,1736312287279-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2025-01-08T04:58:16,571 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T04:58:16,590 WARN [Thread-397 {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T04:58:17,002 INFO [Thread-397 {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2025-01-08T04:58:17,009 INFO [Thread-397 {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2025-01-08T04:58:17,009 INFO [Thread-397 {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2025-01-08T04:58:17,011 INFO [Thread-397 {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2025-01-08T04:58:17,015 WARN [Time-limited test {}] servlet.GuiceFilter(102): Multiple Servlet injectors detected. This is a warning indicating that you have more than one GuiceFilter running in your web application. If this is deliberate, you may safely ignore this message. If this is NOT deliberate however, your application may not work as expected. 2025-01-08T04:58:17,016 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2025-01-08T04:58:17,019 INFO [Thread-397 {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@2c84573a{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop.log.dir/,AVAILABLE} 2025-01-08T04:58:17,020 INFO [Thread-397 {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@5c5cd02{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/static,AVAILABLE} 2025-01-08T04:58:17,055 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2025-01-08T04:58:17,055 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2025-01-08T04:58:17,055 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2025-01-08T04:58:17,073 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T04:58:17,085 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@379c9ae7{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop.log.dir/,AVAILABLE} 2025-01-08T04:58:17,086 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3022421d{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/static,AVAILABLE} 2025-01-08T04:58:17,217 INFO [Thread-397 {}] container.GuiceComponentProviderFactory(113): Registering org.apache.hadoop.yarn.server.resourcemanager.webapp.JAXBContextResolver as a provider class 2025-01-08T04:58:17,218 INFO [Thread-397 {}] container.GuiceComponentProviderFactory(116): Registering org.apache.hadoop.yarn.server.resourcemanager.webapp.RMWebServices as a root resource class 2025-01-08T04:58:17,218 INFO [Thread-397 {}] container.GuiceComponentProviderFactory(113): Registering org.apache.hadoop.yarn.webapp.GenericExceptionHandler as a provider class 2025-01-08T04:58:17,220 INFO [Thread-397 {}] application.WebApplicationImpl(815): Initiating Jersey application, version 'Jersey: 1.19.4 05/24/2017 03:20 PM' 2025-01-08T04:58:17,320 INFO [Thread-397 {}] container.GuiceComponentProviderFactory(168): Binding org.apache.hadoop.yarn.server.resourcemanager.webapp.JAXBContextResolver to GuiceManagedComponentProvider with the scope "Singleton" 2025-01-08T04:58:17,589 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T04:58:17,682 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2025-01-08T04:58:17,683 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:namespace' 2025-01-08T04:58:17,685 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:acl' 2025-01-08T04:58:17,861 INFO [Thread-397 {}] container.GuiceComponentProviderFactory(168): Binding org.apache.hadoop.yarn.webapp.GenericExceptionHandler to GuiceManagedComponentProvider with the scope "Singleton" 2025-01-08T04:58:18,345 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T04:58:18,346 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.security.access.AccessController Metrics about HBase RegionObservers 2025-01-08T04:58:18,347 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_master_table_store 2025-01-08T04:58:18,347 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_master_table_store Metrics about Tables on a single HBase RegionServer 2025-01-08T04:58:18,349 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=Coprocessor.RegionServer.CP_org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T04:58:18,349 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=Coprocessor.RegionServer.CP_org.apache.hadoop.hbase.security.access.AccessController Metrics about HBase RegionServerObservers 2025-01-08T04:58:18,349 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_meta 2025-01-08T04:58:18,349 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_meta Metrics about Tables on a single HBase RegionServer 2025-01-08T04:58:18,352 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_namespace 2025-01-08T04:58:18,352 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_namespace Metrics about Tables on a single HBase RegionServer 2025-01-08T04:58:18,354 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_acl 2025-01-08T04:58:18,354 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_acl Metrics about Tables on a single HBase RegionServer 2025-01-08T04:58:18,356 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: Master,sub=Coprocessor.Master.CP_org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T04:58:18,356 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering Master,sub=Coprocessor.Master.CP_org.apache.hadoop.hbase.security.access.AccessController Metrics about HBase MasterObservers 2025-01-08T04:58:18,356 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: Master,sub=Coprocessor.Master.CP_org.apache.hadoop.hbase.security.access.SecureTestUtil$MasterSyncObserver 2025-01-08T04:58:18,356 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering Master,sub=Coprocessor.Master.CP_org.apache.hadoop.hbase.security.access.SecureTestUtil$MasterSyncObserver Metrics about HBase MasterObservers 2025-01-08T04:58:18,357 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2025-01-08T04:58:18,357 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint Metrics about HBase RegionObservers 2025-01-08T04:58:18,423 INFO [Thread-397 {}] container.GuiceComponentProviderFactory(168): Binding org.apache.hadoop.yarn.server.resourcemanager.webapp.RMWebServices to GuiceManagedComponentProvider with the scope "Singleton" 2025-01-08T04:58:18,455 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@1a80546{cluster,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/java.io.tmpdir/jetty-localhost-44241-hadoop-yarn-common-3_4_1_jar-_-any-14596496996141523183/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/cluster} 2025-01-08T04:58:18,455 INFO [Thread-397 {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@6d9e8f32{jobhistory,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/java.io.tmpdir/jetty-localhost-39749-hadoop-yarn-common-3_4_1_jar-_-any-1785408166838898493/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/jobhistory} 2025-01-08T04:58:18,456 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@31f02c44{HTTP/1.1, (http/1.1)}{localhost:44241} 2025-01-08T04:58:18,456 INFO [Time-limited test {}] server.Server(415): Started @21674ms 2025-01-08T04:58:18,464 INFO [Thread-397 {}] server.AbstractConnector(333): Started ServerConnector@1b10f19e{HTTP/1.1, (http/1.1)}{localhost:39749} 2025-01-08T04:58:18,464 INFO [Thread-397 {}] server.Server(415): Started @21681ms 2025-01-08T04:58:18,806 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741843_1019 (size=5) 2025-01-08T04:58:18,808 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741843_1019 (size=5) 2025-01-08T04:58:18,809 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741843_1019 (size=5) 2025-01-08T04:58:20,856 WARN [Time-limited test {}] tracker.NMLogAggregationStatusTracker(95): Log Aggregation is disabled.So is the LogAggregationStatusTracker. 2025-01-08T04:58:20,863 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T04:58:20,951 WARN [Time-limited test {}] servlet.GuiceFilter(102): Multiple Servlet injectors detected. This is a warning indicating that you have more than one GuiceFilter running in your web application. If this is deliberate, you may safely ignore this message. If this is NOT deliberate however, your application may not work as expected. 2025-01-08T04:58:20,952 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2025-01-08T04:58:21,038 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2025-01-08T04:58:21,038 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2025-01-08T04:58:21,038 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2025-01-08T04:58:21,052 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T04:58:21,058 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@34de0adc{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop.log.dir/,AVAILABLE} 2025-01-08T04:58:21,059 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@23dc62be{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/static,AVAILABLE} 2025-01-08T04:58:21,210 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(116): Registering org.apache.hadoop.yarn.server.nodemanager.webapp.NMWebServices as a root resource class 2025-01-08T04:58:21,210 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(113): Registering org.apache.hadoop.yarn.webapp.GenericExceptionHandler as a provider class 2025-01-08T04:58:21,211 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(113): Registering org.apache.hadoop.yarn.server.nodemanager.webapp.JAXBContextResolver as a provider class 2025-01-08T04:58:21,211 INFO [Time-limited test {}] application.WebApplicationImpl(815): Initiating Jersey application, version 'Jersey: 1.19.4 05/24/2017 03:20 PM' 2025-01-08T04:58:21,248 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(168): Binding org.apache.hadoop.yarn.server.nodemanager.webapp.JAXBContextResolver to GuiceManagedComponentProvider with the scope "Singleton" 2025-01-08T04:58:21,319 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(168): Binding org.apache.hadoop.yarn.webapp.GenericExceptionHandler to GuiceManagedComponentProvider with the scope "Singleton" 2025-01-08T04:58:21,548 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(168): Binding org.apache.hadoop.yarn.server.nodemanager.webapp.NMWebServices to GuiceManagedComponentProvider with the scope "Singleton" 2025-01-08T04:58:21,566 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@56fd86b7{node,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/java.io.tmpdir/jetty-localhost-41649-hadoop-yarn-common-3_4_1_jar-_-any-5281860389259907524/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/node} 2025-01-08T04:58:21,568 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@22c3ce0e{HTTP/1.1, (http/1.1)}{localhost:41649} 2025-01-08T04:58:21,568 INFO [Time-limited test {}] server.Server(415): Started @24785ms 2025-01-08T04:58:22,145 WARN [Time-limited test {}] tracker.NMLogAggregationStatusTracker(95): Log Aggregation is disabled.So is the LogAggregationStatusTracker. 2025-01-08T04:58:22,148 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T04:58:22,180 WARN [Time-limited test {}] servlet.GuiceFilter(102): Multiple Servlet injectors detected. This is a warning indicating that you have more than one GuiceFilter running in your web application. If this is deliberate, you may safely ignore this message. If this is NOT deliberate however, your application may not work as expected. 2025-01-08T04:58:22,181 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2025-01-08T04:58:22,199 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2025-01-08T04:58:22,199 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2025-01-08T04:58:22,199 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2025-01-08T04:58:22,214 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T04:58:22,218 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@4578ed9f{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop.log.dir/,AVAILABLE} 2025-01-08T04:58:22,219 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@7da597ef{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/static,AVAILABLE} 2025-01-08T04:58:22,371 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(116): Registering org.apache.hadoop.yarn.server.nodemanager.webapp.NMWebServices as a root resource class 2025-01-08T04:58:22,371 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(113): Registering org.apache.hadoop.yarn.webapp.GenericExceptionHandler as a provider class 2025-01-08T04:58:22,371 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(113): Registering org.apache.hadoop.yarn.server.nodemanager.webapp.JAXBContextResolver as a provider class 2025-01-08T04:58:22,371 INFO [Time-limited test {}] application.WebApplicationImpl(815): Initiating Jersey application, version 'Jersey: 1.19.4 05/24/2017 03:20 PM' 2025-01-08T04:58:22,407 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(168): Binding org.apache.hadoop.yarn.server.nodemanager.webapp.JAXBContextResolver to GuiceManagedComponentProvider with the scope "Singleton" 2025-01-08T04:58:22,424 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(168): Binding org.apache.hadoop.yarn.webapp.GenericExceptionHandler to GuiceManagedComponentProvider with the scope "Singleton" 2025-01-08T04:58:22,769 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(168): Binding org.apache.hadoop.yarn.server.nodemanager.webapp.NMWebServices to GuiceManagedComponentProvider with the scope "Singleton" 2025-01-08T04:58:22,808 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@451087f8{node,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/java.io.tmpdir/jetty-localhost-33477-hadoop-yarn-common-3_4_1_jar-_-any-7818553130973127097/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/node} 2025-01-08T04:58:22,835 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@306d4cf6{HTTP/1.1, (http/1.1)}{localhost:33477} 2025-01-08T04:58:22,837 INFO [Time-limited test {}] server.Server(415): Started @26054ms 2025-01-08T04:58:22,931 INFO [Time-limited test {}] hbase.HBaseTestingUtility(2825): Mini mapreduce cluster started 2025-01-08T04:58:22,933 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [30,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T04:58:23,041 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: snapshot.TestMobSecureExportSnapshot#testExportWithTargetName Thread=714, OpenFileDescriptor=773, MaxFileDescriptor=1048576, SystemLoadAverage=1190, ProcessCount=21, AvailableMemoryMB=1387 2025-01-08T04:58:23,041 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=714 is superior to 500 2025-01-08T04:58:23,071 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2025-01-08T04:58:23,092 INFO [RS-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:49930, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2025-01-08T04:58:23,107 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.3 create 'testtb-testExportWithTargetName', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T04:58:23,115 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=12, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testExportWithTargetName 2025-01-08T04:58:23,123 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=12, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportWithTargetName execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T04:58:23,126 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=12, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testExportWithTargetName execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T04:58:23,126 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "testtb-testExportWithTargetName" procId is: 12 2025-01-08T04:58:23,139 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=12 2025-01-08T04:58:23,215 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741844_1020 (size=442) 2025-01-08T04:58:23,216 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741844_1020 (size=442) 2025-01-08T04:58:23,216 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741844_1020 (size=442) 2025-01-08T04:58:23,228 INFO [RegionOpenAndInit-testtb-testExportWithTargetName-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => b37060327adf5c369982f72fa6addccc, NAME => 'testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testtb-testExportWithTargetName', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T04:58:23,232 INFO [RegionOpenAndInit-testtb-testExportWithTargetName-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => 002c4d7e8e4b179f821a8b6ed2e4d21a, NAME => 'testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testtb-testExportWithTargetName', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T04:58:23,242 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=12 2025-01-08T04:58:23,356 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741845_1021 (size=67) 2025-01-08T04:58:23,356 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741845_1021 (size=67) 2025-01-08T04:58:23,359 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T04:58:23,359 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741845_1021 (size=67) 2025-01-08T04:58:23,360 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-0 {}] regionserver.HRegion(1681): Closing b37060327adf5c369982f72fa6addccc, disabling compactions & flushes 2025-01-08T04:58:23,360 INFO [RegionOpenAndInit-testtb-testExportWithTargetName-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc. 2025-01-08T04:58:23,360 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc. 2025-01-08T04:58:23,360 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc. after waiting 0 ms 2025-01-08T04:58:23,360 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc. 2025-01-08T04:58:23,360 INFO [RegionOpenAndInit-testtb-testExportWithTargetName-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc. 2025-01-08T04:58:23,360 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-0 {}] regionserver.HRegion(1635): Region close journal for b37060327adf5c369982f72fa6addccc: 2025-01-08T04:58:23,385 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741846_1022 (size=67) 2025-01-08T04:58:23,386 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741846_1022 (size=67) 2025-01-08T04:58:23,387 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741846_1022 (size=67) 2025-01-08T04:58:23,388 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T04:58:23,388 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-1 {}] regionserver.HRegion(1681): Closing 002c4d7e8e4b179f821a8b6ed2e4d21a, disabling compactions & flushes 2025-01-08T04:58:23,388 INFO [RegionOpenAndInit-testtb-testExportWithTargetName-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a. 2025-01-08T04:58:23,388 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a. 2025-01-08T04:58:23,388 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a. after waiting 0 ms 2025-01-08T04:58:23,388 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a. 2025-01-08T04:58:23,388 INFO [RegionOpenAndInit-testtb-testExportWithTargetName-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a. 2025-01-08T04:58:23,388 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-1 {}] regionserver.HRegion(1635): Region close journal for 002c4d7e8e4b179f821a8b6ed2e4d21a: 2025-01-08T04:58:23,391 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=12, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testExportWithTargetName execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T04:58:23,391 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc.","families":{"info":[{"qualifier":"regioninfo","vlen":66,"tag":[],"timestamp":"1736312303391"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736312303391"}]},"ts":"1736312303391"} 2025-01-08T04:58:23,392 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a.","families":{"info":[{"qualifier":"regioninfo","vlen":66,"tag":[],"timestamp":"1736312303391"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736312303391"}]},"ts":"1736312303391"} 2025-01-08T04:58:23,452 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=12 2025-01-08T04:58:23,465 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T04:58:23,471 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=12, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testExportWithTargetName execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T04:58:23,471 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithTargetName","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312303471"}]},"ts":"1736312303471"} 2025-01-08T04:58:23,477 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithTargetName, state=ENABLING in hbase:meta 2025-01-08T04:58:23,484 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(202): Hosts are {f28094b72178=0} racks are {/default-rack=0} 2025-01-08T04:58:23,489 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T04:58:23,489 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T04:58:23,489 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T04:58:23,489 INFO [PEWorker-4 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T04:58:23,489 INFO [PEWorker-4 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T04:58:23,489 INFO [PEWorker-4 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T04:58:23,489 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T04:58:23,490 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=13, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=b37060327adf5c369982f72fa6addccc, ASSIGN}, {pid=14, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=002c4d7e8e4b179f821a8b6ed2e4d21a, ASSIGN}] 2025-01-08T04:58:23,515 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=14, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=002c4d7e8e4b179f821a8b6ed2e4d21a, ASSIGN 2025-01-08T04:58:23,515 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=13, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=b37060327adf5c369982f72fa6addccc, ASSIGN 2025-01-08T04:58:23,517 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=14, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=002c4d7e8e4b179f821a8b6ed2e4d21a, ASSIGN; state=OFFLINE, location=f28094b72178,37583,1736312288873; forceNewPlan=false, retain=false 2025-01-08T04:58:23,517 INFO [PEWorker-3 {}] assignment.TransitRegionStateProcedure(264): Starting pid=13, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=b37060327adf5c369982f72fa6addccc, ASSIGN; state=OFFLINE, location=f28094b72178,37967,1736312288941; forceNewPlan=false, retain=false 2025-01-08T04:58:23,668 INFO [f28094b72178:42195 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T04:58:23,668 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=14 updating hbase:meta row=002c4d7e8e4b179f821a8b6ed2e4d21a, regionState=OPENING, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T04:58:23,668 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=13 updating hbase:meta row=b37060327adf5c369982f72fa6addccc, regionState=OPENING, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T04:58:23,676 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=15, ppid=13, state=RUNNABLE; OpenRegionProcedure b37060327adf5c369982f72fa6addccc, server=f28094b72178,37967,1736312288941}] 2025-01-08T04:58:23,678 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=16, ppid=14, state=RUNNABLE; OpenRegionProcedure 002c4d7e8e4b179f821a8b6ed2e4d21a, server=f28094b72178,37583,1736312288873}] 2025-01-08T04:58:23,755 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=12 2025-01-08T04:58:23,833 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T04:58:23,834 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2025-01-08T04:58:23,841 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T04:58:23,862 INFO [RS-EventLoopGroup-5-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:55364, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2025-01-08T04:58:23,870 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] handler.AssignRegionHandler(135): Open testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a. 2025-01-08T04:58:23,870 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(7285): Opening region: {ENCODED => 002c4d7e8e4b179f821a8b6ed2e4d21a, NAME => 'testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T04:58:23,871 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a. service=AccessControlService 2025-01-08T04:58:23,871 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T04:58:23,872 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportWithTargetName 002c4d7e8e4b179f821a8b6ed2e4d21a 2025-01-08T04:58:23,872 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(894): Instantiated testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T04:58:23,872 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(7327): checking encryption for 002c4d7e8e4b179f821a8b6ed2e4d21a 2025-01-08T04:58:23,872 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(7330): checking classloading for 002c4d7e8e4b179f821a8b6ed2e4d21a 2025-01-08T04:58:23,883 INFO [StoreOpener-002c4d7e8e4b179f821a8b6ed2e4d21a-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 002c4d7e8e4b179f821a8b6ed2e4d21a 2025-01-08T04:58:23,887 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] handler.AssignRegionHandler(135): Open testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc. 2025-01-08T04:58:23,887 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(7285): Opening region: {ENCODED => b37060327adf5c369982f72fa6addccc, NAME => 'testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T04:58:23,888 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc. service=AccessControlService 2025-01-08T04:58:23,888 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T04:58:23,888 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportWithTargetName b37060327adf5c369982f72fa6addccc 2025-01-08T04:58:23,889 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(894): Instantiated testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T04:58:23,889 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(7327): checking encryption for b37060327adf5c369982f72fa6addccc 2025-01-08T04:58:23,889 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(7330): checking classloading for b37060327adf5c369982f72fa6addccc 2025-01-08T04:58:23,893 INFO [StoreOpener-002c4d7e8e4b179f821a8b6ed2e4d21a-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 002c4d7e8e4b179f821a8b6ed2e4d21a columnFamilyName cf 2025-01-08T04:58:23,895 INFO [StoreOpener-b37060327adf5c369982f72fa6addccc-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region b37060327adf5c369982f72fa6addccc 2025-01-08T04:58:23,898 INFO [StoreOpener-b37060327adf5c369982f72fa6addccc-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region b37060327adf5c369982f72fa6addccc columnFamilyName cf 2025-01-08T04:58:23,899 DEBUG [StoreOpener-002c4d7e8e4b179f821a8b6ed2e4d21a-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:58:23,901 INFO [StoreOpener-002c4d7e8e4b179f821a8b6ed2e4d21a-1 {}] regionserver.HStore(327): Store=002c4d7e8e4b179f821a8b6ed2e4d21a/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T04:58:23,901 DEBUG [StoreOpener-b37060327adf5c369982f72fa6addccc-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:58:23,903 INFO [StoreOpener-b37060327adf5c369982f72fa6addccc-1 {}] regionserver.HStore(327): Store=b37060327adf5c369982f72fa6addccc/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T04:58:23,903 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/002c4d7e8e4b179f821a8b6ed2e4d21a 2025-01-08T04:58:23,904 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/002c4d7e8e4b179f821a8b6ed2e4d21a 2025-01-08T04:58:23,907 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/b37060327adf5c369982f72fa6addccc 2025-01-08T04:58:23,908 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/b37060327adf5c369982f72fa6addccc 2025-01-08T04:58:23,910 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(1085): writing seq id for 002c4d7e8e4b179f821a8b6ed2e4d21a 2025-01-08T04:58:23,914 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(1085): writing seq id for b37060327adf5c369982f72fa6addccc 2025-01-08T04:58:23,918 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/002c4d7e8e4b179f821a8b6ed2e4d21a/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T04:58:23,919 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(1102): Opened 002c4d7e8e4b179f821a8b6ed2e4d21a; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=69014428, jitterRate=0.028395116329193115}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T04:58:23,920 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/b37060327adf5c369982f72fa6addccc/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T04:58:23,921 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(1001): Region open journal for 002c4d7e8e4b179f821a8b6ed2e4d21a: 2025-01-08T04:58:23,921 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(1102): Opened b37060327adf5c369982f72fa6addccc; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=66157014, jitterRate=-0.014183670282363892}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T04:58:23,921 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(1001): Region open journal for b37060327adf5c369982f72fa6addccc: 2025-01-08T04:58:23,922 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a., pid=16, masterSystemTime=1736312303840 2025-01-08T04:58:23,925 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a. 2025-01-08T04:58:23,925 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] handler.AssignRegionHandler(164): Opened testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a. 2025-01-08T04:58:23,926 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc., pid=15, masterSystemTime=1736312303833 2025-01-08T04:58:23,931 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=14 updating hbase:meta row=002c4d7e8e4b179f821a8b6ed2e4d21a, regionState=OPEN, openSeqNum=2, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T04:58:23,932 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc. 2025-01-08T04:58:23,932 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] handler.AssignRegionHandler(164): Opened testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc. 2025-01-08T04:58:23,935 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=13 updating hbase:meta row=b37060327adf5c369982f72fa6addccc, regionState=OPEN, openSeqNum=2, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T04:58:23,942 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=16, resume processing ppid=14 2025-01-08T04:58:23,942 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=16, ppid=14, state=SUCCESS; OpenRegionProcedure 002c4d7e8e4b179f821a8b6ed2e4d21a, server=f28094b72178,37583,1736312288873 in 260 msec 2025-01-08T04:58:23,946 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=15, resume processing ppid=13 2025-01-08T04:58:23,946 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=14, ppid=12, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=002c4d7e8e4b179f821a8b6ed2e4d21a, ASSIGN in 452 msec 2025-01-08T04:58:23,946 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=15, ppid=13, state=SUCCESS; OpenRegionProcedure b37060327adf5c369982f72fa6addccc, server=f28094b72178,37967,1736312288941 in 265 msec 2025-01-08T04:58:23,951 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=13, resume processing ppid=12 2025-01-08T04:58:23,951 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=13, ppid=12, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=b37060327adf5c369982f72fa6addccc, ASSIGN in 456 msec 2025-01-08T04:58:23,952 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=12, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testExportWithTargetName execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T04:58:23,952 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithTargetName","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312303952"}]},"ts":"1736312303952"} 2025-01-08T04:58:23,955 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithTargetName, state=ENABLED in hbase:meta 2025-01-08T04:58:23,960 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=12, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportWithTargetName execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T04:58:23,965 DEBUG [PEWorker-3 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testExportWithTargetName jenkins: RWXCA 2025-01-08T04:58:23,983 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37583 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportWithTargetName], kv [jenkins: RWXCA] 2025-01-08T04:58:24,014 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/acl 2025-01-08T04:58:24,014 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/acl 2025-01-08T04:58:24,014 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/acl 2025-01-08T04:58:24,015 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:24,015 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/acl 2025-01-08T04:58:24,015 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:24,015 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:24,015 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T04:58:24,033 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithTargetName with data PBUF\x0AE\x0A\x07jenkins\x12:\x08\x03"6\x0A*\x0A\x07default\x12\x1Ftesttb-testExportWithTargetName \x00 \x01 \x02 \x03 \x04 2025-01-08T04:58:24,033 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithTargetName with data PBUF\x0AE\x0A\x07jenkins\x12:\x08\x03"6\x0A*\x0A\x07default\x12\x1Ftesttb-testExportWithTargetName \x00 \x01 \x02 \x03 \x04 2025-01-08T04:58:24,034 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithTargetName with data PBUF\x0AE\x0A\x07jenkins\x12:\x08\x03"6\x0A*\x0A\x07default\x12\x1Ftesttb-testExportWithTargetName \x00 \x01 \x02 \x03 \x04 2025-01-08T04:58:24,034 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithTargetName with data PBUF\x0AE\x0A\x07jenkins\x12:\x08\x03"6\x0A*\x0A\x07default\x12\x1Ftesttb-testExportWithTargetName \x00 \x01 \x02 \x03 \x04 2025-01-08T04:58:24,043 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=12, state=SUCCESS; CreateTableProcedure table=testtb-testExportWithTargetName in 927 msec 2025-01-08T04:58:24,257 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=12 2025-01-08T04:58:24,258 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testExportWithTargetName, procId: 12 completed 2025-01-08T04:58:24,263 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportWithTargetName 2025-01-08T04:58:24,265 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc. 2025-01-08T04:58:24,266 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T04:58:24,306 DEBUG [hconnection-0x39b30dcc-shared-pool-4 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T04:58:24,311 INFO [RS-EventLoopGroup-5-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:55370, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T04:58:24,342 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } 2025-01-08T04:58:24,342 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736312304342 (current time:1736312304342). 2025-01-08T04:58:24,342 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T04:58:24,342 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot emptySnaptb0-testExportWithTargetName VERSION not specified, setting to 2 2025-01-08T04:58:24,343 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T04:58:24,344 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x06cf96cb to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@3c3c3714 2025-01-08T04:58:24,351 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2b44253b, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T04:58:24,357 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T04:58:24,359 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:49942, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T04:58:24,362 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x06cf96cb to 127.0.0.1:59048 2025-01-08T04:58:24,362 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T04:58:24,366 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x56485e4f to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@17c36a46 2025-01-08T04:58:24,379 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@6024430a, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T04:58:24,393 DEBUG [hconnection-0x49d9e751-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T04:58:24,395 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:49954, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T04:58:24,401 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x56485e4f to 127.0.0.1:59048 2025-01-08T04:58:24,401 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T04:58:24,401 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportWithTargetName], kv [jenkins: RWXCA] 2025-01-08T04:58:24,425 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T04:58:24,436 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=17, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } 2025-01-08T04:58:24,437 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 }, snapshot procedure id = 17 2025-01-08T04:58:24,438 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T04:58:24,444 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=17 2025-01-08T04:58:24,444 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T04:58:24,462 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T04:58:24,514 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741847_1023 (size=167) 2025-01-08T04:58:24,515 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741847_1023 (size=167) 2025-01-08T04:58:24,517 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741847_1023 (size=167) 2025-01-08T04:58:24,525 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T04:58:24,528 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=18, ppid=17, state=RUNNABLE; SnapshotRegionProcedure b37060327adf5c369982f72fa6addccc}, {pid=19, ppid=17, state=RUNNABLE; SnapshotRegionProcedure 002c4d7e8e4b179f821a8b6ed2e4d21a}] 2025-01-08T04:58:24,536 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=18, ppid=17, state=RUNNABLE; SnapshotRegionProcedure b37060327adf5c369982f72fa6addccc 2025-01-08T04:58:24,536 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=19, ppid=17, state=RUNNABLE; SnapshotRegionProcedure 002c4d7e8e4b179f821a8b6ed2e4d21a 2025-01-08T04:58:24,546 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=17 2025-01-08T04:58:24,696 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T04:58:24,696 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T04:58:24,698 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37967 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=18 2025-01-08T04:58:24,698 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=37583 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=19 2025-01-08T04:58:24,703 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=19}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a. 2025-01-08T04:58:24,706 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=19}] regionserver.HRegion(2538): Flush status journal for 002c4d7e8e4b179f821a8b6ed2e4d21a: 2025-01-08T04:58:24,706 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=18}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc. 2025-01-08T04:58:24,706 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=19}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a. for emptySnaptb0-testExportWithTargetName completed. 2025-01-08T04:58:24,707 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=18}] regionserver.HRegion(2538): Flush status journal for b37060327adf5c369982f72fa6addccc: 2025-01-08T04:58:24,707 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=18}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc. for emptySnaptb0-testExportWithTargetName completed. 2025-01-08T04:58:24,708 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=18}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc.' region-info for snapshot=emptySnaptb0-testExportWithTargetName 2025-01-08T04:58:24,708 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=19}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a.' region-info for snapshot=emptySnaptb0-testExportWithTargetName 2025-01-08T04:58:24,712 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=19}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T04:58:24,713 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=18}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T04:58:24,716 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=19}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T04:58:24,718 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=18}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T04:58:24,748 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=17 2025-01-08T04:58:24,811 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741849_1025 (size=70) 2025-01-08T04:58:24,813 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741849_1025 (size=70) 2025-01-08T04:58:24,814 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741849_1025 (size=70) 2025-01-08T04:58:24,819 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=18}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc. 2025-01-08T04:58:24,822 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=18}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=18 2025-01-08T04:58:24,825 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=18 2025-01-08T04:58:24,826 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportWithTargetName on region b37060327adf5c369982f72fa6addccc 2025-01-08T04:58:24,826 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=18, ppid=17, state=RUNNABLE; SnapshotRegionProcedure b37060327adf5c369982f72fa6addccc 2025-01-08T04:58:24,833 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=18, ppid=17, state=SUCCESS; SnapshotRegionProcedure b37060327adf5c369982f72fa6addccc in 300 msec 2025-01-08T04:58:24,840 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741848_1024 (size=70) 2025-01-08T04:58:24,843 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741848_1024 (size=70) 2025-01-08T04:58:24,843 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741848_1024 (size=70) 2025-01-08T04:58:24,845 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=19}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a. 2025-01-08T04:58:24,845 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=19}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=19 2025-01-08T04:58:24,846 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=19 2025-01-08T04:58:24,847 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportWithTargetName on region 002c4d7e8e4b179f821a8b6ed2e4d21a 2025-01-08T04:58:24,847 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=19, ppid=17, state=RUNNABLE; SnapshotRegionProcedure 002c4d7e8e4b179f821a8b6ed2e4d21a 2025-01-08T04:58:24,867 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=19, resume processing ppid=17 2025-01-08T04:58:24,867 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=19, ppid=17, state=SUCCESS; SnapshotRegionProcedure 002c4d7e8e4b179f821a8b6ed2e4d21a in 325 msec 2025-01-08T04:58:24,867 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T04:58:24,870 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T04:58:24,886 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T04:58:24,887 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T04:58:24,888 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:58:24,889 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(328): No files under family: cf 2025-01-08T04:58:24,990 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741850_1026 (size=62) 2025-01-08T04:58:24,991 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741850_1026 (size=62) 2025-01-08T04:58:24,991 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741850_1026 (size=62) 2025-01-08T04:58:25,051 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=17 2025-01-08T04:58:25,394 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T04:58:25,395 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for emptySnaptb0-testExportWithTargetName 2025-01-08T04:58:25,398 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/emptySnaptb0-testExportWithTargetName 2025-01-08T04:58:25,555 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=17 2025-01-08T04:58:25,564 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741851_1027 (size=649) 2025-01-08T04:58:25,569 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741851_1027 (size=649) 2025-01-08T04:58:25,570 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741851_1027 (size=649) 2025-01-08T04:58:25,607 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T04:58:25,667 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T04:58:25,675 DEBUG [PEWorker-3 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/emptySnaptb0-testExportWithTargetName to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/emptySnaptb0-testExportWithTargetName 2025-01-08T04:58:25,691 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T04:58:25,692 DEBUG [PEWorker-3 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 }, snapshot procedure id = 17 2025-01-08T04:58:25,703 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=17, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } in 1.2620 sec 2025-01-08T04:58:26,557 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=17 2025-01-08T04:58:26,558 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportWithTargetName, procId: 17 completed 2025-01-08T04:58:26,584 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37583 {}] regionserver.HRegion(8254): writing data to region testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T04:58:26,600 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37967 {}] regionserver.HRegion(8254): writing data to region testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T04:58:26,608 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportWithTargetName 2025-01-08T04:58:26,608 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc. 2025-01-08T04:58:26,608 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T04:58:26,641 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } 2025-01-08T04:58:26,641 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736312306641 (current time:1736312306641). 2025-01-08T04:58:26,641 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T04:58:26,641 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testExportWithTargetName VERSION not specified, setting to 2 2025-01-08T04:58:26,641 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T04:58:26,643 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x7081d456 to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@6072c609 2025-01-08T04:58:26,671 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5599ae17, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T04:58:26,677 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T04:58:26,679 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:45558, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T04:58:26,681 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x7081d456 to 127.0.0.1:59048 2025-01-08T04:58:26,681 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T04:58:26,683 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x3442e0da to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@42762ad5 2025-01-08T04:58:26,729 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5a07dd9a, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T04:58:26,752 DEBUG [hconnection-0x6d5aca07-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T04:58:26,754 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:45566, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T04:58:26,759 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x3442e0da to 127.0.0.1:59048 2025-01-08T04:58:26,760 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T04:58:26,760 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportWithTargetName], kv [jenkins: RWXCA] 2025-01-08T04:58:26,763 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T04:58:26,766 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=20, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } 2025-01-08T04:58:26,768 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 }, snapshot procedure id = 20 2025-01-08T04:58:26,770 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=20 2025-01-08T04:58:26,772 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T04:58:26,778 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T04:58:26,785 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T04:58:26,872 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=20 2025-01-08T04:58:26,909 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741852_1028 (size=162) 2025-01-08T04:58:26,914 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741852_1028 (size=162) 2025-01-08T04:58:26,920 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T04:58:26,920 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=21, ppid=20, state=RUNNABLE; SnapshotRegionProcedure b37060327adf5c369982f72fa6addccc}, {pid=22, ppid=20, state=RUNNABLE; SnapshotRegionProcedure 002c4d7e8e4b179f821a8b6ed2e4d21a}] 2025-01-08T04:58:26,921 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741852_1028 (size=162) 2025-01-08T04:58:26,925 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=22, ppid=20, state=RUNNABLE; SnapshotRegionProcedure 002c4d7e8e4b179f821a8b6ed2e4d21a 2025-01-08T04:58:26,925 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=21, ppid=20, state=RUNNABLE; SnapshotRegionProcedure b37060327adf5c369982f72fa6addccc 2025-01-08T04:58:27,076 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=20 2025-01-08T04:58:27,086 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T04:58:27,086 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T04:58:27,087 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=37583 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=22 2025-01-08T04:58:27,087 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=37967 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=21 2025-01-08T04:58:27,097 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc. 2025-01-08T04:58:27,098 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.HRegion(2837): Flushing b37060327adf5c369982f72fa6addccc 1/1 column families, dataSize=199 B heapSize=688 B 2025-01-08T04:58:27,117 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a. 2025-01-08T04:58:27,128 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.HRegion(2837): Flushing 002c4d7e8e4b179f821a8b6ed2e4d21a 1/1 column families, dataSize=3.06 KB heapSize=6.86 KB 2025-01-08T04:58:27,380 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=20 2025-01-08T04:58:27,425 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e202501083d787e24ea294d52811dd5bc303d5061_b37060327adf5c369982f72fa6addccc is 71, key is 03b0afb64f8ff0c8e55ddb65ca9b7c75/cf:q/1736312306600/Put/seqid=0 2025-01-08T04:58:27,449 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b202501088bc02a47cccd476eb432b6014a73cde4_002c4d7e8e4b179f821a8b6ed2e4d21a is 71, key is 10b13d225bb997bd2ed00af91b5691c2/cf:q/1736312306584/Put/seqid=0 2025-01-08T04:58:27,558 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741854_1030 (size=5101) 2025-01-08T04:58:27,564 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:58:27,565 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741854_1030 (size=5101) 2025-01-08T04:58:27,565 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741854_1030 (size=5101) 2025-01-08T04:58:27,601 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741853_1029 (size=8171) 2025-01-08T04:58:27,604 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741853_1029 (size=8171) 2025-01-08T04:58:27,605 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741853_1029 (size=8171) 2025-01-08T04:58:27,606 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:58:27,750 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b202501088bc02a47cccd476eb432b6014a73cde4_002c4d7e8e4b179f821a8b6ed2e4d21a to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/c4ca4238a0b923820dcc509a6f75849b202501088bc02a47cccd476eb432b6014a73cde4_002c4d7e8e4b179f821a8b6ed2e4d21a 2025-01-08T04:58:27,750 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e202501083d787e24ea294d52811dd5bc303d5061_b37060327adf5c369982f72fa6addccc to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/d41d8cd98f00b204e9800998ecf8427e202501083d787e24ea294d52811dd5bc303d5061_b37060327adf5c369982f72fa6addccc 2025-01-08T04:58:27,752 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/002c4d7e8e4b179f821a8b6ed2e4d21a/.tmp/cf/53fa94fd5eab484fa318fe20bfff2138, store: [table=testtb-testExportWithTargetName family=cf region=002c4d7e8e4b179f821a8b6ed2e4d21a] 2025-01-08T04:58:27,755 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/b37060327adf5c369982f72fa6addccc/.tmp/cf/e5d3aa2dc4084501becdaf876a425a07, store: [table=testtb-testExportWithTargetName family=cf region=b37060327adf5c369982f72fa6addccc] 2025-01-08T04:58:27,775 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/b37060327adf5c369982f72fa6addccc/.tmp/cf/e5d3aa2dc4084501becdaf876a425a07 is 208, key is 0764610999e94fc4a648d119f59efe484/cf:q/1736312306600/Put/seqid=0 2025-01-08T04:58:27,781 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/002c4d7e8e4b179f821a8b6ed2e4d21a/.tmp/cf/53fa94fd5eab484fa318fe20bfff2138 is 208, key is 13ddf6b111c0a23caea52a2a4b64fd124/cf:q/1736312306584/Put/seqid=0 2025-01-08T04:58:27,819 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741855_1031 (size=5912) 2025-01-08T04:58:27,820 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741855_1031 (size=5912) 2025-01-08T04:58:27,820 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741855_1031 (size=5912) 2025-01-08T04:58:27,820 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741856_1032 (size=14949) 2025-01-08T04:58:27,821 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=199, hasBloomFilter=true, into tmp file hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/b37060327adf5c369982f72fa6addccc/.tmp/cf/e5d3aa2dc4084501becdaf876a425a07 2025-01-08T04:58:27,828 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741856_1032 (size=14949) 2025-01-08T04:58:27,829 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741856_1032 (size=14949) 2025-01-08T04:58:27,832 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=3.1 K, hasBloomFilter=true, into tmp file hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/002c4d7e8e4b179f821a8b6ed2e4d21a/.tmp/cf/53fa94fd5eab484fa318fe20bfff2138 2025-01-08T04:58:27,851 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/b37060327adf5c369982f72fa6addccc/.tmp/cf/e5d3aa2dc4084501becdaf876a425a07 as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/b37060327adf5c369982f72fa6addccc/cf/e5d3aa2dc4084501becdaf876a425a07 2025-01-08T04:58:27,853 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/002c4d7e8e4b179f821a8b6ed2e4d21a/.tmp/cf/53fa94fd5eab484fa318fe20bfff2138 as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/002c4d7e8e4b179f821a8b6ed2e4d21a/cf/53fa94fd5eab484fa318fe20bfff2138 2025-01-08T04:58:27,871 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/b37060327adf5c369982f72fa6addccc/cf/e5d3aa2dc4084501becdaf876a425a07, entries=3, sequenceid=6, filesize=5.8 K 2025-01-08T04:58:27,875 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.HRegion(3040): Finished flush of dataSize ~199 B/199, heapSize ~672 B/672, currentSize=0 B/0 for b37060327adf5c369982f72fa6addccc in 777ms, sequenceid=6, compaction requested=false 2025-01-08T04:58:27,876 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportWithTargetName' 2025-01-08T04:58:27,877 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.HRegion(2538): Flush status journal for b37060327adf5c369982f72fa6addccc: 2025-01-08T04:58:27,877 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc. for snaptb0-testExportWithTargetName completed. 2025-01-08T04:58:27,879 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc.' region-info for snapshot=snaptb0-testExportWithTargetName 2025-01-08T04:58:27,879 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T04:58:27,879 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/b37060327adf5c369982f72fa6addccc/cf/e5d3aa2dc4084501becdaf876a425a07] hfiles 2025-01-08T04:58:27,879 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/b37060327adf5c369982f72fa6addccc/cf/e5d3aa2dc4084501becdaf876a425a07 for snapshot=snaptb0-testExportWithTargetName 2025-01-08T04:58:27,888 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/002c4d7e8e4b179f821a8b6ed2e4d21a/cf/53fa94fd5eab484fa318fe20bfff2138, entries=47, sequenceid=6, filesize=14.6 K 2025-01-08T04:58:27,894 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.HRegion(3040): Finished flush of dataSize ~3.06 KB/3137, heapSize ~6.84 KB/7008, currentSize=0 B/0 for 002c4d7e8e4b179f821a8b6ed2e4d21a in 775ms, sequenceid=6, compaction requested=false 2025-01-08T04:58:27,894 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.HRegion(2538): Flush status journal for 002c4d7e8e4b179f821a8b6ed2e4d21a: 2025-01-08T04:58:27,894 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a. for snaptb0-testExportWithTargetName completed. 2025-01-08T04:58:27,895 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a.' region-info for snapshot=snaptb0-testExportWithTargetName 2025-01-08T04:58:27,895 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T04:58:27,895 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/002c4d7e8e4b179f821a8b6ed2e4d21a/cf/53fa94fd5eab484fa318fe20bfff2138] hfiles 2025-01-08T04:58:27,895 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/002c4d7e8e4b179f821a8b6ed2e4d21a/cf/53fa94fd5eab484fa318fe20bfff2138 for snapshot=snaptb0-testExportWithTargetName 2025-01-08T04:58:27,899 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=20 2025-01-08T04:58:27,969 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741857_1033 (size=109) 2025-01-08T04:58:27,969 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741857_1033 (size=109) 2025-01-08T04:58:27,970 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc. 2025-01-08T04:58:27,970 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741857_1033 (size=109) 2025-01-08T04:58:27,970 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=21 2025-01-08T04:58:27,971 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=21 2025-01-08T04:58:27,972 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportWithTargetName on region b37060327adf5c369982f72fa6addccc 2025-01-08T04:58:27,972 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=21, ppid=20, state=RUNNABLE; SnapshotRegionProcedure b37060327adf5c369982f72fa6addccc 2025-01-08T04:58:27,975 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741858_1034 (size=109) 2025-01-08T04:58:27,976 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741858_1034 (size=109) 2025-01-08T04:58:27,978 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=21, ppid=20, state=SUCCESS; SnapshotRegionProcedure b37060327adf5c369982f72fa6addccc in 1.0540 sec 2025-01-08T04:58:27,978 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741858_1034 (size=109) 2025-01-08T04:58:27,978 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a. 2025-01-08T04:58:27,978 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=22 2025-01-08T04:58:27,979 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=22 2025-01-08T04:58:27,979 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportWithTargetName on region 002c4d7e8e4b179f821a8b6ed2e4d21a 2025-01-08T04:58:27,982 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=22, ppid=20, state=RUNNABLE; SnapshotRegionProcedure 002c4d7e8e4b179f821a8b6ed2e4d21a 2025-01-08T04:58:27,987 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=22, resume processing ppid=20 2025-01-08T04:58:27,987 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=22, ppid=20, state=SUCCESS; SnapshotRegionProcedure 002c4d7e8e4b179f821a8b6ed2e4d21a in 1.0640 sec 2025-01-08T04:58:27,987 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T04:58:27,988 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T04:58:27,990 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T04:58:27,990 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T04:58:27,990 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:58:27,997 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/c4ca4238a0b923820dcc509a6f75849b202501088bc02a47cccd476eb432b6014a73cde4_002c4d7e8e4b179f821a8b6ed2e4d21a, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/d41d8cd98f00b204e9800998ecf8427e202501083d787e24ea294d52811dd5bc303d5061_b37060327adf5c369982f72fa6addccc] hfiles 2025-01-08T04:58:27,998 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/c4ca4238a0b923820dcc509a6f75849b202501088bc02a47cccd476eb432b6014a73cde4_002c4d7e8e4b179f821a8b6ed2e4d21a 2025-01-08T04:58:27,998 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/d41d8cd98f00b204e9800998ecf8427e202501083d787e24ea294d52811dd5bc303d5061_b37060327adf5c369982f72fa6addccc 2025-01-08T04:58:28,037 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741859_1035 (size=293) 2025-01-08T04:58:28,039 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741859_1035 (size=293) 2025-01-08T04:58:28,039 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741859_1035 (size=293) 2025-01-08T04:58:28,051 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T04:58:28,052 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testExportWithTargetName 2025-01-08T04:58:28,054 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/snaptb0-testExportWithTargetName 2025-01-08T04:58:28,129 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741860_1036 (size=959) 2025-01-08T04:58:28,131 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741860_1036 (size=959) 2025-01-08T04:58:28,131 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741860_1036 (size=959) 2025-01-08T04:58:28,156 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T04:58:28,199 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T04:58:28,200 DEBUG [PEWorker-1 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/snaptb0-testExportWithTargetName to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportWithTargetName 2025-01-08T04:58:28,207 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T04:58:28,207 DEBUG [PEWorker-1 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 }, snapshot procedure id = 20 2025-01-08T04:58:28,210 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=20, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } in 1.4440 sec 2025-01-08T04:58:28,346 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportWithTargetName 2025-01-08T04:58:28,347 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportWithTargetName Metrics about Tables on a single HBase RegionServer 2025-01-08T04:58:28,906 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=20 2025-01-08T04:58:28,914 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportWithTargetName, procId: 20 completed 2025-01-08T04:58:28,915 INFO [Time-limited test {}] snapshot.TestExportSnapshot(468): HDFS export destination path: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312308915 2025-01-08T04:58:28,915 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=hdfs://localhost:45939, tgtDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312308915, rawTgtDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312308915, srcFsUri=hdfs://localhost:45939, srcDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T04:58:29,019 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:45939, inputRoot=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T04:58:29,019 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1149249371_22, ugi=jenkins (auth:SIMPLE)]], outputRoot=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312308915, skipTmp=false, initialOutputSnapshotDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312308915/.hbase-snapshot/.tmp/testExportWithTargetName 2025-01-08T04:58:29,031 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T04:58:29,072 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportWithTargetName to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312308915/.hbase-snapshot/.tmp/testExportWithTargetName 2025-01-08T04:58:29,140 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741861_1037 (size=959) 2025-01-08T04:58:29,141 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741861_1037 (size=959) 2025-01-08T04:58:29,144 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741861_1037 (size=959) 2025-01-08T04:58:29,173 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741862_1038 (size=162) 2025-01-08T04:58:29,173 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741862_1038 (size=162) 2025-01-08T04:58:29,173 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741862_1038 (size=162) 2025-01-08T04:58:29,487 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T04:58:29,588 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741863_1039 (size=154) 2025-01-08T04:58:29,588 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741863_1039 (size=154) 2025-01-08T04:58:29,588 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741863_1039 (size=154) 2025-01-08T04:58:29,597 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.7.0-SNAPSHOT.jar 2025-01-08T04:58:29,598 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T04:58:29,598 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T04:58:29,599 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T04:58:30,870 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop-15830700566538635148.jar 2025-01-08T04:58:30,870 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T04:58:30,871 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T04:58:30,959 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop-10926367082166630586.jar 2025-01-08T04:58:30,959 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T04:58:30,960 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T04:58:30,961 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T04:58:30,961 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T04:58:30,962 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T04:58:30,962 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T04:58:30,963 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T04:58:30,963 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T04:58:30,964 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T04:58:30,964 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T04:58:30,965 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T04:58:30,965 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T04:58:30,966 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T04:58:30,966 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T04:58:30,966 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T04:58:30,967 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T04:58:30,967 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T04:58:30,968 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T04:58:30,971 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T04:58:30,971 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T04:58:30,972 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T04:58:30,972 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T04:58:30,973 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T04:58:30,974 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T04:58:30,975 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T04:58:31,251 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741864_1040 (size=127628) 2025-01-08T04:58:31,252 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741864_1040 (size=127628) 2025-01-08T04:58:31,252 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741864_1040 (size=127628) 2025-01-08T04:58:31,377 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741865_1041 (size=2172137) 2025-01-08T04:58:31,379 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741865_1041 (size=2172137) 2025-01-08T04:58:31,379 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741865_1041 (size=2172137) 2025-01-08T04:58:31,467 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741866_1042 (size=451756) 2025-01-08T04:58:31,471 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741866_1042 (size=451756) 2025-01-08T04:58:31,471 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741866_1042 (size=451756) 2025-01-08T04:58:31,598 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741867_1043 (size=213228) 2025-01-08T04:58:31,598 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741867_1043 (size=213228) 2025-01-08T04:58:31,600 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741867_1043 (size=213228) 2025-01-08T04:58:31,799 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741868_1044 (size=6350712) 2025-01-08T04:58:31,799 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741868_1044 (size=6350712) 2025-01-08T04:58:31,800 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741868_1044 (size=6350712) 2025-01-08T04:58:32,286 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741869_1045 (size=1877034) 2025-01-08T04:58:32,286 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741869_1045 (size=1877034) 2025-01-08T04:58:32,290 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741869_1045 (size=1877034) 2025-01-08T04:58:32,366 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741870_1046 (size=533455) 2025-01-08T04:58:32,367 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741870_1046 (size=533455) 2025-01-08T04:58:32,368 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741870_1046 (size=533455) 2025-01-08T04:58:32,526 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741871_1047 (size=7280644) 2025-01-08T04:58:32,527 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741871_1047 (size=7280644) 2025-01-08T04:58:32,527 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741871_1047 (size=7280644) 2025-01-08T04:58:32,618 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741872_1048 (size=4188619) 2025-01-08T04:58:32,623 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741872_1048 (size=4188619) 2025-01-08T04:58:32,624 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741872_1048 (size=4188619) 2025-01-08T04:58:32,716 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741873_1049 (size=20406) 2025-01-08T04:58:32,716 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741873_1049 (size=20406) 2025-01-08T04:58:32,717 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741873_1049 (size=20406) 2025-01-08T04:58:32,757 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741874_1050 (size=75495) 2025-01-08T04:58:32,758 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741874_1050 (size=75495) 2025-01-08T04:58:32,758 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741874_1050 (size=75495) 2025-01-08T04:58:32,801 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741875_1051 (size=45609) 2025-01-08T04:58:32,802 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741875_1051 (size=45609) 2025-01-08T04:58:32,802 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741875_1051 (size=45609) 2025-01-08T04:58:32,837 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741876_1052 (size=110084) 2025-01-08T04:58:32,838 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741876_1052 (size=110084) 2025-01-08T04:58:32,839 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741876_1052 (size=110084) 2025-01-08T04:58:32,895 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741877_1053 (size=1323991) 2025-01-08T04:58:32,898 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741877_1053 (size=1323991) 2025-01-08T04:58:32,902 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741877_1053 (size=1323991) 2025-01-08T04:58:32,965 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741878_1054 (size=23076) 2025-01-08T04:58:32,967 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741878_1054 (size=23076) 2025-01-08T04:58:32,967 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741878_1054 (size=23076) 2025-01-08T04:58:33,017 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741879_1055 (size=126803) 2025-01-08T04:58:33,021 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741879_1055 (size=126803) 2025-01-08T04:58:33,022 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741879_1055 (size=126803) 2025-01-08T04:58:33,070 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741880_1056 (size=322274) 2025-01-08T04:58:33,072 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741880_1056 (size=322274) 2025-01-08T04:58:33,073 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741880_1056 (size=322274) 2025-01-08T04:58:33,120 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741881_1057 (size=1832290) 2025-01-08T04:58:33,121 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741881_1057 (size=1832290) 2025-01-08T04:58:33,121 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741881_1057 (size=1832290) 2025-01-08T04:58:33,147 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741882_1058 (size=30081) 2025-01-08T04:58:33,150 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741882_1058 (size=30081) 2025-01-08T04:58:33,150 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741882_1058 (size=30081) 2025-01-08T04:58:33,176 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741883_1059 (size=53616) 2025-01-08T04:58:33,176 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741883_1059 (size=53616) 2025-01-08T04:58:33,179 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741883_1059 (size=53616) 2025-01-08T04:58:33,201 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741884_1060 (size=29229) 2025-01-08T04:58:33,206 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741884_1060 (size=29229) 2025-01-08T04:58:33,210 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741884_1060 (size=29229) 2025-01-08T04:58:33,236 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741885_1061 (size=169089) 2025-01-08T04:58:33,236 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741885_1061 (size=169089) 2025-01-08T04:58:33,236 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741885_1061 (size=169089) 2025-01-08T04:58:33,753 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741886_1062 (size=5175431) 2025-01-08T04:58:33,753 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741886_1062 (size=5175431) 2025-01-08T04:58:33,754 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741886_1062 (size=5175431) 2025-01-08T04:58:33,867 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741887_1063 (size=136454) 2025-01-08T04:58:33,868 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741887_1063 (size=136454) 2025-01-08T04:58:33,868 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741887_1063 (size=136454) 2025-01-08T04:58:34,324 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741888_1064 (size=907470) 2025-01-08T04:58:34,325 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741888_1064 (size=907470) 2025-01-08T04:58:34,330 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741888_1064 (size=907470) 2025-01-08T04:58:34,407 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741889_1065 (size=3317408) 2025-01-08T04:58:34,411 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741889_1065 (size=3317408) 2025-01-08T04:58:34,411 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741889_1065 (size=3317408) 2025-01-08T04:58:34,464 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741890_1066 (size=503880) 2025-01-08T04:58:34,465 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741890_1066 (size=503880) 2025-01-08T04:58:34,466 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741890_1066 (size=503880) 2025-01-08T04:58:34,587 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741891_1067 (size=4695811) 2025-01-08T04:58:34,587 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741891_1067 (size=4695811) 2025-01-08T04:58:34,587 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741891_1067 (size=4695811) 2025-01-08T04:58:34,999 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T04:58:35,028 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'snaptb0-testExportWithTargetName' hfile list 2025-01-08T04:58:35,061 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(751): export split=0 size=33.3 K 2025-01-08T04:58:35,195 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741892_1068 (size=722) 2025-01-08T04:58:35,196 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741892_1068 (size=722) 2025-01-08T04:58:35,196 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741892_1068 (size=722) 2025-01-08T04:58:35,295 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741893_1069 (size=15) 2025-01-08T04:58:35,295 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741893_1069 (size=15) 2025-01-08T04:58:35,296 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741893_1069 (size=15) 2025-01-08T04:58:35,786 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T04:58:36,026 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741894_1070 (size=304886) 2025-01-08T04:58:36,027 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741894_1070 (size=304886) 2025-01-08T04:58:36,027 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741894_1070 (size=304886) 2025-01-08T04:58:37,280 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T04:58:37,280 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T04:58:37,532 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0001_000001 (auth:SIMPLE) from 127.0.0.1:49464 2025-01-08T04:58:40,767 DEBUG [BootstrapNodeManager {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2025-01-08T04:58:40,769 INFO [RS-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:59922, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.1 (auth:SIMPLE), service=RegionServerStatusService 2025-01-08T04:58:43,783 DEBUG [BootstrapNodeManager {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2025-01-08T04:58:43,785 INFO [RS-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:59928, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.2 (auth:SIMPLE), service=RegionServerStatusService 2025-01-08T04:58:44,767 DEBUG [BootstrapNodeManager {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2025-01-08T04:58:44,772 INFO [RS-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:54210, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=RegionServerStatusService 2025-01-08T04:58:51,303 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0001_000001 (auth:SIMPLE) from 127.0.0.1:39856 2025-01-08T04:58:52,027 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741895_1071 (size=350560) 2025-01-08T04:58:52,030 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741895_1071 (size=350560) 2025-01-08T04:58:52,030 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741895_1071 (size=350560) 2025-01-08T04:58:53,849 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0001_000001 (auth:SIMPLE) from 127.0.0.1:59212 2025-01-08T04:58:54,835 INFO [master/f28094b72178:0.Chore.1 {}] migrate.RollingUpgradeChore(116): There is no table to migrate StoreFileTracker! 2025-01-08T04:58:54,835 INFO [master/f28094b72178:0.Chore.1 {}] migrate.RollingUpgradeChore(85): All Rolling-Upgrade tasks are complete, shutdown RollingUpgradeChore! 2025-01-08T04:58:59,886 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741896_1072 (size=14949) 2025-01-08T04:58:59,890 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741896_1072 (size=14949) 2025-01-08T04:58:59,891 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741896_1072 (size=14949) 2025-01-08T04:58:59,970 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741897_1073 (size=8171) 2025-01-08T04:58:59,970 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741897_1073 (size=8171) 2025-01-08T04:58:59,970 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741897_1073 (size=8171) 2025-01-08T04:59:00,032 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741898_1074 (size=5912) 2025-01-08T04:59:00,032 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741898_1074 (size=5912) 2025-01-08T04:59:00,032 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741898_1074 (size=5912) 2025-01-08T04:59:00,092 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741899_1075 (size=5101) 2025-01-08T04:59:00,092 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741899_1075 (size=5101) 2025-01-08T04:59:00,093 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741899_1075 (size=5101) 2025-01-08T04:59:00,198 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741900_1076 (size=17461) 2025-01-08T04:59:00,199 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741900_1076 (size=17461) 2025-01-08T04:59:00,203 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741900_1076 (size=17461) 2025-01-08T04:59:00,221 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741901_1077 (size=464) 2025-01-08T04:59:00,221 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741901_1077 (size=464) 2025-01-08T04:59:00,222 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741901_1077 (size=464) 2025-01-08T04:59:00,310 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741902_1078 (size=17461) 2025-01-08T04:59:00,310 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741902_1078 (size=17461) 2025-01-08T04:59:00,311 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741902_1078 (size=17461) 2025-01-08T04:59:00,369 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741903_1079 (size=350560) 2025-01-08T04:59:00,369 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741903_1079 (size=350560) 2025-01-08T04:59:00,370 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741903_1079 (size=350560) 2025-01-08T04:59:00,383 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_2/usercache/jenkins/appcache/application_1736312298603_0001/container_1736312298603_0001_01_000002/launch_container.sh] 2025-01-08T04:59:00,383 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_2/usercache/jenkins/appcache/application_1736312298603_0001/container_1736312298603_0001_01_000002/container_tokens] 2025-01-08T04:59:00,384 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_2/usercache/jenkins/appcache/application_1736312298603_0001/container_1736312298603_0001_01_000002/sysfs] 2025-01-08T04:59:00,388 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0001_000001 (auth:SIMPLE) from 127.0.0.1:43950 2025-01-08T04:59:01,672 INFO [Time-limited test {}] snapshot.ExportSnapshot(1207): Finalize the Snapshot Export 2025-01-08T04:59:01,706 INFO [Time-limited test {}] snapshot.ExportSnapshot(1218): Verify the exported snapshot's expiration status and integrity. 2025-01-08T04:59:01,757 INFO [Time-limited test {}] snapshot.ExportSnapshot(1224): Export Completed: testExportWithTargetName 2025-01-08T04:59:01,757 INFO [Time-limited test {}] snapshot.TestExportSnapshot(362): Exported snapshot 2025-01-08T04:59:01,758 INFO [Time-limited test {}] snapshot.TestExportSnapshot(373): Verified filesystem state 2025-01-08T04:59:01,758 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1149249371_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportWithTargetName at hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportWithTargetName 2025-01-08T04:59:01,759 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportWithTargetName/.snapshotinfo 2025-01-08T04:59:01,759 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportWithTargetName/data.manifest 2025-01-08T04:59:01,759 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1149249371_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312308915/.hbase-snapshot/testExportWithTargetName at hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312308915/.hbase-snapshot/testExportWithTargetName 2025-01-08T04:59:01,764 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312308915/.hbase-snapshot/testExportWithTargetName/.snapshotinfo 2025-01-08T04:59:01,764 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312308915/.hbase-snapshot/testExportWithTargetName/data.manifest 2025-01-08T04:59:01,806 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testExportWithTargetName 2025-01-08T04:59:01,811 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.3 disable testtb-testExportWithTargetName 2025-01-08T04:59:01,823 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=23, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testExportWithTargetName 2025-01-08T04:59:01,851 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=23 2025-01-08T04:59:01,859 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithTargetName","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312341859"}]},"ts":"1736312341859"} 2025-01-08T04:59:01,866 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithTargetName, state=DISABLING in hbase:meta 2025-01-08T04:59:01,881 INFO [PEWorker-2 {}] procedure.DisableTableProcedure(284): Set testtb-testExportWithTargetName to state=DISABLING 2025-01-08T04:59:01,884 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=24, ppid=23, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testExportWithTargetName}] 2025-01-08T04:59:01,896 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=25, ppid=24, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=b37060327adf5c369982f72fa6addccc, UNASSIGN}, {pid=26, ppid=24, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=002c4d7e8e4b179f821a8b6ed2e4d21a, UNASSIGN}] 2025-01-08T04:59:01,898 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=26, ppid=24, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=002c4d7e8e4b179f821a8b6ed2e4d21a, UNASSIGN 2025-01-08T04:59:01,898 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=25, ppid=24, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=b37060327adf5c369982f72fa6addccc, UNASSIGN 2025-01-08T04:59:01,900 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=26 updating hbase:meta row=002c4d7e8e4b179f821a8b6ed2e4d21a, regionState=CLOSING, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T04:59:01,900 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=25 updating hbase:meta row=b37060327adf5c369982f72fa6addccc, regionState=CLOSING, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T04:59:01,904 DEBUG [PEWorker-5 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T04:59:01,904 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=27, ppid=26, state=RUNNABLE; CloseRegionProcedure 002c4d7e8e4b179f821a8b6ed2e4d21a, server=f28094b72178,37583,1736312288873}] 2025-01-08T04:59:01,910 DEBUG [PEWorker-3 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T04:59:01,910 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=28, ppid=25, state=RUNNABLE; CloseRegionProcedure b37060327adf5c369982f72fa6addccc, server=f28094b72178,37967,1736312288941}] 2025-01-08T04:59:01,953 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=23 2025-01-08T04:59:02,065 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T04:59:02,065 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T04:59:02,079 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] handler.UnassignRegionHandler(124): Close b37060327adf5c369982f72fa6addccc 2025-01-08T04:59:02,080 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T04:59:02,081 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] regionserver.HRegion(1681): Closing b37060327adf5c369982f72fa6addccc, disabling compactions & flushes 2025-01-08T04:59:02,082 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] regionserver.HRegion(1703): Closing region testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc. 2025-01-08T04:59:02,082 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc. 2025-01-08T04:59:02,082 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc. after waiting 0 ms 2025-01-08T04:59:02,082 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc. 2025-01-08T04:59:02,091 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] handler.UnassignRegionHandler(124): Close 002c4d7e8e4b179f821a8b6ed2e4d21a 2025-01-08T04:59:02,091 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T04:59:02,092 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] regionserver.HRegion(1681): Closing 002c4d7e8e4b179f821a8b6ed2e4d21a, disabling compactions & flushes 2025-01-08T04:59:02,092 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] regionserver.HRegion(1703): Closing region testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a. 2025-01-08T04:59:02,092 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a. 2025-01-08T04:59:02,092 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a. after waiting 0 ms 2025-01-08T04:59:02,092 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a. 2025-01-08T04:59:02,156 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=23 2025-01-08T04:59:02,218 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/b37060327adf5c369982f72fa6addccc/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T04:59:02,219 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/002c4d7e8e4b179f821a8b6ed2e4d21a/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T04:59:02,230 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T04:59:02,230 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T04:59:02,231 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] regionserver.HRegion(1922): Closed testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a. 2025-01-08T04:59:02,231 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] regionserver.HRegion(1635): Region close journal for 002c4d7e8e4b179f821a8b6ed2e4d21a: 2025-01-08T04:59:02,231 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] regionserver.HRegion(1922): Closed testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc. 2025-01-08T04:59:02,231 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] regionserver.HRegion(1635): Region close journal for b37060327adf5c369982f72fa6addccc: 2025-01-08T04:59:02,236 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] handler.UnassignRegionHandler(170): Closed 002c4d7e8e4b179f821a8b6ed2e4d21a 2025-01-08T04:59:02,239 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=26 updating hbase:meta row=002c4d7e8e4b179f821a8b6ed2e4d21a, regionState=CLOSED 2025-01-08T04:59:02,240 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] handler.UnassignRegionHandler(170): Closed b37060327adf5c369982f72fa6addccc 2025-01-08T04:59:02,241 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=25 updating hbase:meta row=b37060327adf5c369982f72fa6addccc, regionState=CLOSED 2025-01-08T04:59:02,253 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=27, resume processing ppid=26 2025-01-08T04:59:02,254 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=28, resume processing ppid=25 2025-01-08T04:59:02,254 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=27, ppid=26, state=SUCCESS; CloseRegionProcedure 002c4d7e8e4b179f821a8b6ed2e4d21a, server=f28094b72178,37583,1736312288873 in 338 msec 2025-01-08T04:59:02,257 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=28, ppid=25, state=SUCCESS; CloseRegionProcedure b37060327adf5c369982f72fa6addccc, server=f28094b72178,37967,1736312288941 in 335 msec 2025-01-08T04:59:02,261 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=26, ppid=24, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=002c4d7e8e4b179f821a8b6ed2e4d21a, UNASSIGN in 357 msec 2025-01-08T04:59:02,273 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=25, resume processing ppid=24 2025-01-08T04:59:02,273 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=25, ppid=24, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=b37060327adf5c369982f72fa6addccc, UNASSIGN in 361 msec 2025-01-08T04:59:02,287 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=24, resume processing ppid=23 2025-01-08T04:59:02,287 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=24, ppid=23, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testExportWithTargetName in 398 msec 2025-01-08T04:59:02,290 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithTargetName","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312342290"}]},"ts":"1736312342290"} 2025-01-08T04:59:02,292 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithTargetName, state=DISABLED in hbase:meta 2025-01-08T04:59:02,299 INFO [PEWorker-4 {}] procedure.DisableTableProcedure(296): Set testtb-testExportWithTargetName to state=DISABLED 2025-01-08T04:59:02,304 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=23, state=SUCCESS; DisableTableProcedure table=testtb-testExportWithTargetName in 488 msec 2025-01-08T04:59:02,457 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=23 2025-01-08T04:59:02,458 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testExportWithTargetName, procId: 23 completed 2025-01-08T04:59:02,462 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.3 delete testtb-testExportWithTargetName 2025-01-08T04:59:02,471 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=29, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testExportWithTargetName 2025-01-08T04:59:02,477 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=29, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportWithTargetName 2025-01-08T04:59:02,479 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testExportWithTargetName 2025-01-08T04:59:02,479 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=29, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testExportWithTargetName 2025-01-08T04:59:02,484 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37583 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testExportWithTargetName 2025-01-08T04:59:02,487 DEBUG [HFileArchiver-1 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/b37060327adf5c369982f72fa6addccc 2025-01-08T04:59:02,490 DEBUG [HFileArchiver-2 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/002c4d7e8e4b179f821a8b6ed2e4d21a 2025-01-08T04:59:02,494 DEBUG [HFileArchiver-2 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/002c4d7e8e4b179f821a8b6ed2e4d21a/cf, FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/002c4d7e8e4b179f821a8b6ed2e4d21a/recovered.edits] 2025-01-08T04:59:02,496 DEBUG [HFileArchiver-1 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/b37060327adf5c369982f72fa6addccc/cf, FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/b37060327adf5c369982f72fa6addccc/recovered.edits] 2025-01-08T04:59:02,514 DEBUG [HFileArchiver-2 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/002c4d7e8e4b179f821a8b6ed2e4d21a/cf/53fa94fd5eab484fa318fe20bfff2138 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportWithTargetName/002c4d7e8e4b179f821a8b6ed2e4d21a/cf/53fa94fd5eab484fa318fe20bfff2138 2025-01-08T04:59:02,514 DEBUG [HFileArchiver-1 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/b37060327adf5c369982f72fa6addccc/cf/e5d3aa2dc4084501becdaf876a425a07 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportWithTargetName/b37060327adf5c369982f72fa6addccc/cf/e5d3aa2dc4084501becdaf876a425a07 2025-01-08T04:59:02,520 DEBUG [HFileArchiver-1 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/b37060327adf5c369982f72fa6addccc/recovered.edits/9.seqid to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportWithTargetName/b37060327adf5c369982f72fa6addccc/recovered.edits/9.seqid 2025-01-08T04:59:02,521 DEBUG [HFileArchiver-2 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/002c4d7e8e4b179f821a8b6ed2e4d21a/recovered.edits/9.seqid to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportWithTargetName/002c4d7e8e4b179f821a8b6ed2e4d21a/recovered.edits/9.seqid 2025-01-08T04:59:02,521 DEBUG [HFileArchiver-1 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/b37060327adf5c369982f72fa6addccc 2025-01-08T04:59:02,522 DEBUG [HFileArchiver-2 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithTargetName/002c4d7e8e4b179f821a8b6ed2e4d21a 2025-01-08T04:59:02,522 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(313): Archived testtb-testExportWithTargetName regions 2025-01-08T04:59:02,523 DEBUG [PEWorker-2 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71 2025-01-08T04:59:02,525 DEBUG [PEWorker-2 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf] 2025-01-08T04:59:02,533 DEBUG [PEWorker-2 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/c4ca4238a0b923820dcc509a6f75849b202501088bc02a47cccd476eb432b6014a73cde4_002c4d7e8e4b179f821a8b6ed2e4d21a to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/c4ca4238a0b923820dcc509a6f75849b202501088bc02a47cccd476eb432b6014a73cde4_002c4d7e8e4b179f821a8b6ed2e4d21a 2025-01-08T04:59:02,536 DEBUG [PEWorker-2 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/d41d8cd98f00b204e9800998ecf8427e202501083d787e24ea294d52811dd5bc303d5061_b37060327adf5c369982f72fa6addccc to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/d41d8cd98f00b204e9800998ecf8427e202501083d787e24ea294d52811dd5bc303d5061_b37060327adf5c369982f72fa6addccc 2025-01-08T04:59:02,537 DEBUG [PEWorker-2 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71 2025-01-08T04:59:02,541 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=29, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testExportWithTargetName 2025-01-08T04:59:02,546 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithTargetName 2025-01-08T04:59:02,546 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithTargetName 2025-01-08T04:59:02,546 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithTargetName 2025-01-08T04:59:02,546 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithTargetName 2025-01-08T04:59:02,547 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=37583 {}] util.ReflectedFunctionCache(97): Populated cache for org.apache.hadoop.hbase.filter.KeyOnlyFilter in 0ms 2025-01-08T04:59:02,549 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithTargetName with data PBUF 2025-01-08T04:59:02,549 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithTargetName with data PBUF 2025-01-08T04:59:02,549 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithTargetName with data PBUF 2025-01-08T04:59:02,552 WARN [PEWorker-2 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testtb-testExportWithTargetName from hbase:meta 2025-01-08T04:59:02,559 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testExportWithTargetName' descriptor. 2025-01-08T04:59:02,563 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=29, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testExportWithTargetName 2025-01-08T04:59:02,563 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testExportWithTargetName' from region states. 2025-01-08T04:59:02,564 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736312342564"}]},"ts":"9223372036854775807"} 2025-01-08T04:59:02,564 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736312342564"}]},"ts":"9223372036854775807"} 2025-01-08T04:59:02,569 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportWithTargetName 2025-01-08T04:59:02,569 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportWithTargetName 2025-01-08T04:59:02,569 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T04:59:02,569 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportWithTargetName 2025-01-08T04:59:02,569 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T04:59:02,569 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T04:59:02,571 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T04:59:02,574 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithTargetName with data null 2025-01-08T04:59:02,574 INFO [zk-permission-watcher-pool-0 {}] access.AuthManager(135): Skipping permission cache refresh because writable data is empty 2025-01-08T04:59:02,575 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T04:59:02,575 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => b37060327adf5c369982f72fa6addccc, NAME => 'testtb-testExportWithTargetName,,1736312303106.b37060327adf5c369982f72fa6addccc.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => 002c4d7e8e4b179f821a8b6ed2e4d21a, NAME => 'testtb-testExportWithTargetName,1,1736312303106.002c4d7e8e4b179f821a8b6ed2e4d21a.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T04:59:02,575 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testExportWithTargetName' as deleted. 2025-01-08T04:59:02,575 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportWithTargetName","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736312342575"}]},"ts":"9223372036854775807"} 2025-01-08T04:59:02,577 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=29 2025-01-08T04:59:02,588 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testExportWithTargetName state from META 2025-01-08T04:59:02,623 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(133): Finished pid=29, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportWithTargetName 2025-01-08T04:59:02,626 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=29, state=SUCCESS; DeleteTableProcedure table=testtb-testExportWithTargetName in 159 msec 2025-01-08T04:59:02,685 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=29 2025-01-08T04:59:02,685 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testExportWithTargetName, procId: 29 completed 2025-01-08T04:59:02,714 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "emptySnaptb0-testExportWithTargetName" 2025-01-08T04:59:02,718 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(380): Deleting snapshot: emptySnaptb0-testExportWithTargetName 2025-01-08T04:59:02,723 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "snaptb0-testExportWithTargetName" 2025-01-08T04:59:02,731 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testExportWithTargetName 2025-01-08T04:59:02,773 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: snapshot.TestMobSecureExportSnapshot#testExportWithTargetName Thread=768 (was 714) Potentially hanging thread: ContainersLauncher #0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:401) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_CLOSE_REGION-regionserver/f28094b72178:0-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (366384994) connection to localhost/127.0.0.1:41307 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: hconnection-0x168697cd-shared-pool-5 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ForkJoinPool.commonPool-worker-4 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkUntil(LockSupport.java:410) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1726) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Potentially hanging thread: htable-pool-1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:401) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1149249371_22 at /127.0.0.1:39114 [Waiting for operation #6] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DeletionService #2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Async disk worker #0 for volume /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: zk-permission-watcher-pool-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (366384994) connection to localhost/127.0.0.1:45939 from jenkins.hfs.2 java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: RPCClient-NioEventLoopGroup-6-2 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:62) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:883) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x168697cd-shared-pool-8 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_1902851478_1 at /127.0.0.1:38296 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: process reaper (pid 7231) java.base@17.0.11/java.lang.ProcessHandleImpl.waitForProcessExit0(Native Method) java.base@17.0.11/java.lang.ProcessHandleImpl$1.run(ProcessHandleImpl.java:150) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1149249371_22 at /127.0.0.1:60098 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1149249371_22 at /127.0.0.1:38302 [Waiting for operation #4] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_OPEN_REGION-regionserver/f28094b72178:0-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Async disk worker #0 for volume /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data6 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (366384994) connection to localhost/127.0.0.1:45137 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: hconnection-0x168697cd-shared-pool-7 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_1902851478_1 at /127.0.0.1:39092 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_CLOSE_REGION-regionserver/f28094b72178:0-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Async disk worker #0 for volume /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data5 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x168697cd-shared-pool-6 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ContainersLauncher #0 java.base@17.0.11/java.io.FileInputStream.readBytes(Native Method) java.base@17.0.11/java.io.FileInputStream.read(FileInputStream.java:276) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:282) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) java.base@17.0.11/sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:281) java.base@17.0.11/sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:324) java.base@17.0.11/sun.nio.cs.StreamDecoder.read(StreamDecoder.java:189) java.base@17.0.11/java.io.InputStreamReader.read(InputStreamReader.java:177) java.base@17.0.11/java.io.BufferedReader.fill(BufferedReader.java:162) java.base@17.0.11/java.io.BufferedReader.read1(BufferedReader.java:213) java.base@17.0.11/java.io.BufferedReader.read(BufferedReader.java:287) app//org.apache.hadoop.util.Shell$ShellCommandExecutor.parseExecResult(Shell.java:1295) app//org.apache.hadoop.util.Shell.runCommand(Shell.java:1054) app//org.apache.hadoop.util.Shell.run(Shell.java:959) app//org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:1282) app//org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:349) app//org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.launchContainer(ContainerLaunch.java:600) app//org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:388) app//org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:105) java.base@17.0.11/java.util.concurrent.FutureTask.run(FutureTask.java:264) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DeletionService #1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_OPEN_REGION-regionserver/f28094b72178:0-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-6-3 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:62) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:883) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Async disk worker #0 for volume /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data3 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Thread-1336 java.base@17.0.11/java.io.FileInputStream.readBytes(Native Method) java.base@17.0.11/java.io.FileInputStream.read(FileInputStream.java:276) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:282) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) java.base@17.0.11/sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:281) java.base@17.0.11/sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:324) java.base@17.0.11/sun.nio.cs.StreamDecoder.read(StreamDecoder.java:189) java.base@17.0.11/java.io.InputStreamReader.read(InputStreamReader.java:177) java.base@17.0.11/java.io.BufferedReader.fill(BufferedReader.java:162) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:329) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:396) app//org.apache.hadoop.util.Shell$1.run(Shell.java:1025) Potentially hanging thread: DeletionService #0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:45137 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Async disk worker #0 for volume /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data4 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DeletionService #0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Async disk worker #0 for volume /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:41307 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: zk-permission-watcher-pool-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ForkJoinPool.commonPool-worker-5 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1724) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Potentially hanging thread: htable-pool-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:401) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Container metrics unregistration java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Potentially hanging thread: zk-permission-watcher-pool-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: zk-permission-watcher-pool-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DeletionService #3 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) - Thread LEAK? -, OpenFileDescriptor=805 (was 773) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=1184 (was 1190), ProcessCount=14 (was 21), AvailableMemoryMB=3765 (was 1387) - AvailableMemoryMB LEAK? - 2025-01-08T04:59:02,773 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=768 is superior to 500 2025-01-08T04:59:02,807 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: snapshot.TestMobSecureExportSnapshot#testExportWithResetTtl Thread=768, OpenFileDescriptor=805, MaxFileDescriptor=1048576, SystemLoadAverage=1184, ProcessCount=14, AvailableMemoryMB=3757 2025-01-08T04:59:02,807 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=768 is superior to 500 2025-01-08T04:59:02,810 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.3 create 'testtb-testExportWithResetTtl', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T04:59:02,812 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=30, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testExportWithResetTtl 2025-01-08T04:59:02,813 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=30, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportWithResetTtl execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T04:59:02,813 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "testtb-testExportWithResetTtl" procId is: 30 2025-01-08T04:59:02,814 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=30, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testExportWithResetTtl execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T04:59:02,814 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=30 2025-01-08T04:59:02,857 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741904_1080 (size=440) 2025-01-08T04:59:02,863 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741904_1080 (size=440) 2025-01-08T04:59:02,864 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741904_1080 (size=440) 2025-01-08T04:59:02,871 INFO [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 87d1017402411aba577634d8e1497363, NAME => 'testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testtb-testExportWithResetTtl', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T04:59:02,871 INFO [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => ef5fe42efd5a386cbc2993c0042cbfb9, NAME => 'testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testtb-testExportWithResetTtl', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T04:59:02,916 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741906_1082 (size=65) 2025-01-08T04:59:02,916 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741906_1082 (size=65) 2025-01-08T04:59:02,917 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T04:59:02,917 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1681): Closing ef5fe42efd5a386cbc2993c0042cbfb9, disabling compactions & flushes 2025-01-08T04:59:02,917 INFO [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9. 2025-01-08T04:59:02,917 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9. 2025-01-08T04:59:02,917 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9. after waiting 0 ms 2025-01-08T04:59:02,917 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9. 2025-01-08T04:59:02,917 INFO [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9. 2025-01-08T04:59:02,917 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1635): Region close journal for ef5fe42efd5a386cbc2993c0042cbfb9: 2025-01-08T04:59:02,919 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741906_1082 (size=65) 2025-01-08T04:59:02,920 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741905_1081 (size=65) 2025-01-08T04:59:02,920 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=30 2025-01-08T04:59:02,921 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741905_1081 (size=65) 2025-01-08T04:59:02,922 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T04:59:02,922 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1681): Closing 87d1017402411aba577634d8e1497363, disabling compactions & flushes 2025-01-08T04:59:02,922 INFO [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363. 2025-01-08T04:59:02,922 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363. 2025-01-08T04:59:02,922 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363. after waiting 0 ms 2025-01-08T04:59:02,922 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363. 2025-01-08T04:59:02,922 INFO [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363. 2025-01-08T04:59:02,922 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1635): Region close journal for 87d1017402411aba577634d8e1497363: 2025-01-08T04:59:02,923 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741905_1081 (size=65) 2025-01-08T04:59:02,939 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=30, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testExportWithResetTtl execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T04:59:02,939 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9.","families":{"info":[{"qualifier":"regioninfo","vlen":64,"tag":[],"timestamp":"1736312342939"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736312342939"}]},"ts":"1736312342939"} 2025-01-08T04:59:02,940 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363.","families":{"info":[{"qualifier":"regioninfo","vlen":64,"tag":[],"timestamp":"1736312342939"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736312342939"}]},"ts":"1736312342939"} 2025-01-08T04:59:02,951 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T04:59:02,953 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=30, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testExportWithResetTtl execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T04:59:02,953 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312342953"}]},"ts":"1736312342953"} 2025-01-08T04:59:02,957 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithResetTtl, state=ENABLING in hbase:meta 2025-01-08T04:59:03,092 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(202): Hosts are {f28094b72178=0} racks are {/default-rack=0} 2025-01-08T04:59:03,094 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T04:59:03,094 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T04:59:03,094 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T04:59:03,094 INFO [PEWorker-5 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T04:59:03,094 INFO [PEWorker-5 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T04:59:03,094 INFO [PEWorker-5 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T04:59:03,094 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T04:59:03,095 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=31, ppid=30, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=87d1017402411aba577634d8e1497363, ASSIGN}, {pid=32, ppid=30, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=ef5fe42efd5a386cbc2993c0042cbfb9, ASSIGN}] 2025-01-08T04:59:03,102 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=32, ppid=30, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=ef5fe42efd5a386cbc2993c0042cbfb9, ASSIGN 2025-01-08T04:59:03,102 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=31, ppid=30, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=87d1017402411aba577634d8e1497363, ASSIGN 2025-01-08T04:59:03,107 INFO [PEWorker-5 {}] assignment.TransitRegionStateProcedure(264): Starting pid=31, ppid=30, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=87d1017402411aba577634d8e1497363, ASSIGN; state=OFFLINE, location=f28094b72178,39801,1736312288779; forceNewPlan=false, retain=false 2025-01-08T04:59:03,107 INFO [PEWorker-3 {}] assignment.TransitRegionStateProcedure(264): Starting pid=32, ppid=30, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=ef5fe42efd5a386cbc2993c0042cbfb9, ASSIGN; state=OFFLINE, location=f28094b72178,37583,1736312288873; forceNewPlan=false, retain=false 2025-01-08T04:59:03,122 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=30 2025-01-08T04:59:03,258 INFO [f28094b72178:42195 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T04:59:03,258 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=31 updating hbase:meta row=87d1017402411aba577634d8e1497363, regionState=OPENING, regionLocation=f28094b72178,39801,1736312288779 2025-01-08T04:59:03,261 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=32 updating hbase:meta row=ef5fe42efd5a386cbc2993c0042cbfb9, regionState=OPENING, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T04:59:03,266 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=33, ppid=31, state=RUNNABLE; OpenRegionProcedure 87d1017402411aba577634d8e1497363, server=f28094b72178,39801,1736312288779}] 2025-01-08T04:59:03,268 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=34, ppid=32, state=RUNNABLE; OpenRegionProcedure ef5fe42efd5a386cbc2993c0042cbfb9, server=f28094b72178,37583,1736312288873}] 2025-01-08T04:59:03,419 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,39801,1736312288779 2025-01-08T04:59:03,419 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2025-01-08T04:59:03,421 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T04:59:03,423 INFO [RS-EventLoopGroup-3-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:36806, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2025-01-08T04:59:03,426 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] handler.AssignRegionHandler(135): Open testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9. 2025-01-08T04:59:03,427 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(7285): Opening region: {ENCODED => ef5fe42efd5a386cbc2993c0042cbfb9, NAME => 'testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T04:59:03,427 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=30 2025-01-08T04:59:03,427 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9. service=AccessControlService 2025-01-08T04:59:03,427 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T04:59:03,428 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportWithResetTtl ef5fe42efd5a386cbc2993c0042cbfb9 2025-01-08T04:59:03,428 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(894): Instantiated testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T04:59:03,428 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(7327): checking encryption for ef5fe42efd5a386cbc2993c0042cbfb9 2025-01-08T04:59:03,428 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(7330): checking classloading for ef5fe42efd5a386cbc2993c0042cbfb9 2025-01-08T04:59:03,430 INFO [StoreOpener-ef5fe42efd5a386cbc2993c0042cbfb9-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region ef5fe42efd5a386cbc2993c0042cbfb9 2025-01-08T04:59:03,432 INFO [StoreOpener-ef5fe42efd5a386cbc2993c0042cbfb9-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region ef5fe42efd5a386cbc2993c0042cbfb9 columnFamilyName cf 2025-01-08T04:59:03,433 DEBUG [StoreOpener-ef5fe42efd5a386cbc2993c0042cbfb9-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:59:03,434 INFO [StoreOpener-ef5fe42efd5a386cbc2993c0042cbfb9-1 {}] regionserver.HStore(327): Store=ef5fe42efd5a386cbc2993c0042cbfb9/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T04:59:03,435 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/ef5fe42efd5a386cbc2993c0042cbfb9 2025-01-08T04:59:03,437 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/ef5fe42efd5a386cbc2993c0042cbfb9 2025-01-08T04:59:03,440 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(1085): writing seq id for ef5fe42efd5a386cbc2993c0042cbfb9 2025-01-08T04:59:03,444 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] handler.AssignRegionHandler(135): Open testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363. 2025-01-08T04:59:03,444 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(7285): Opening region: {ENCODED => 87d1017402411aba577634d8e1497363, NAME => 'testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T04:59:03,445 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/ef5fe42efd5a386cbc2993c0042cbfb9/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T04:59:03,445 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(1102): Opened ef5fe42efd5a386cbc2993c0042cbfb9; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=71281952, jitterRate=0.06218385696411133}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T04:59:03,446 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363. service=AccessControlService 2025-01-08T04:59:03,446 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T04:59:03,446 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportWithResetTtl 87d1017402411aba577634d8e1497363 2025-01-08T04:59:03,447 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(894): Instantiated testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T04:59:03,447 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(7327): checking encryption for 87d1017402411aba577634d8e1497363 2025-01-08T04:59:03,447 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(7330): checking classloading for 87d1017402411aba577634d8e1497363 2025-01-08T04:59:03,447 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(1001): Region open journal for ef5fe42efd5a386cbc2993c0042cbfb9: 2025-01-08T04:59:03,448 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9., pid=34, masterSystemTime=1736312343421 2025-01-08T04:59:03,449 INFO [StoreOpener-87d1017402411aba577634d8e1497363-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 87d1017402411aba577634d8e1497363 2025-01-08T04:59:03,452 INFO [StoreOpener-87d1017402411aba577634d8e1497363-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 87d1017402411aba577634d8e1497363 columnFamilyName cf 2025-01-08T04:59:03,453 DEBUG [StoreOpener-87d1017402411aba577634d8e1497363-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:59:03,454 INFO [StoreOpener-87d1017402411aba577634d8e1497363-1 {}] regionserver.HStore(327): Store=87d1017402411aba577634d8e1497363/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T04:59:03,456 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9. 2025-01-08T04:59:03,456 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] handler.AssignRegionHandler(164): Opened testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9. 2025-01-08T04:59:03,456 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/87d1017402411aba577634d8e1497363 2025-01-08T04:59:03,459 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/87d1017402411aba577634d8e1497363 2025-01-08T04:59:03,459 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=32 updating hbase:meta row=ef5fe42efd5a386cbc2993c0042cbfb9, regionState=OPEN, openSeqNum=2, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T04:59:03,463 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(1085): writing seq id for 87d1017402411aba577634d8e1497363 2025-01-08T04:59:03,467 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=34, resume processing ppid=32 2025-01-08T04:59:03,468 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=34, ppid=32, state=SUCCESS; OpenRegionProcedure ef5fe42efd5a386cbc2993c0042cbfb9, server=f28094b72178,37583,1736312288873 in 195 msec 2025-01-08T04:59:03,469 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=32, ppid=30, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=ef5fe42efd5a386cbc2993c0042cbfb9, ASSIGN in 372 msec 2025-01-08T04:59:03,475 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/87d1017402411aba577634d8e1497363/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T04:59:03,476 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(1102): Opened 87d1017402411aba577634d8e1497363; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=72063083, jitterRate=0.07382361590862274}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T04:59:03,477 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(1001): Region open journal for 87d1017402411aba577634d8e1497363: 2025-01-08T04:59:03,478 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363., pid=33, masterSystemTime=1736312343419 2025-01-08T04:59:03,481 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363. 2025-01-08T04:59:03,481 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] handler.AssignRegionHandler(164): Opened testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363. 2025-01-08T04:59:03,482 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=31 updating hbase:meta row=87d1017402411aba577634d8e1497363, regionState=OPEN, openSeqNum=2, regionLocation=f28094b72178,39801,1736312288779 2025-01-08T04:59:03,492 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=33, resume processing ppid=31 2025-01-08T04:59:03,494 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=33, ppid=31, state=SUCCESS; OpenRegionProcedure 87d1017402411aba577634d8e1497363, server=f28094b72178,39801,1736312288779 in 221 msec 2025-01-08T04:59:03,496 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=31, resume processing ppid=30 2025-01-08T04:59:03,497 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=31, ppid=30, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=87d1017402411aba577634d8e1497363, ASSIGN in 397 msec 2025-01-08T04:59:03,497 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=30, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testExportWithResetTtl execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T04:59:03,498 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312343497"}]},"ts":"1736312343497"} 2025-01-08T04:59:03,500 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithResetTtl, state=ENABLED in hbase:meta 2025-01-08T04:59:03,762 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=30, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportWithResetTtl execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T04:59:03,763 DEBUG [PEWorker-2 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testExportWithResetTtl jenkins: RWXCA 2025-01-08T04:59:03,767 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37583 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportWithResetTtl], kv [jenkins: RWXCA] 2025-01-08T04:59:03,863 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T04:59:03,863 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T04:59:03,863 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T04:59:03,863 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T04:59:03,917 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T04:59:03,918 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T04:59:03,918 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T04:59:03,918 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T04:59:03,919 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=30, state=SUCCESS; CreateTableProcedure table=testtb-testExportWithResetTtl in 1.1060 sec 2025-01-08T04:59:03,928 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=30 2025-01-08T04:59:03,929 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testExportWithResetTtl, procId: 30 completed 2025-01-08T04:59:03,948 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportWithResetTtl 2025-01-08T04:59:03,949 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363. 2025-01-08T04:59:03,949 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T04:59:03,989 DEBUG [hconnection-0x39b30dcc-shared-pool-10 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T04:59:03,993 INFO [RS-EventLoopGroup-3-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:36810, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T04:59:04,009 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } 2025-01-08T04:59:04,009 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736312344009 (current time:1736312344009). 2025-01-08T04:59:04,009 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T04:59:04,009 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot emptySnaptb0-testExportWithResetTtl VERSION not specified, setting to 2 2025-01-08T04:59:04,009 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T04:59:04,011 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x251ccb03 to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@32254f82 2025-01-08T04:59:04,085 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@76ae5cf, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T04:59:04,091 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T04:59:04,093 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:50576, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T04:59:04,095 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x251ccb03 to 127.0.0.1:59048 2025-01-08T04:59:04,095 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T04:59:04,097 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x4237cf04 to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@5a76c19f 2025-01-08T04:59:04,208 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@a621383, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T04:59:04,210 DEBUG [hconnection-0x9162df5-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T04:59:04,211 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:50588, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T04:59:04,216 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x4237cf04 to 127.0.0.1:59048 2025-01-08T04:59:04,216 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T04:59:04,216 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportWithResetTtl], kv [jenkins: RWXCA] 2025-01-08T04:59:04,217 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T04:59:04,218 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=35, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } 2025-01-08T04:59:04,218 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 }, snapshot procedure id = 35 2025-01-08T04:59:04,220 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T04:59:04,220 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=35 2025-01-08T04:59:04,221 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T04:59:04,225 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T04:59:04,236 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741907_1083 (size=161) 2025-01-08T04:59:04,237 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741907_1083 (size=161) 2025-01-08T04:59:04,239 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741907_1083 (size=161) 2025-01-08T04:59:04,240 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T04:59:04,241 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=36, ppid=35, state=RUNNABLE; SnapshotRegionProcedure 87d1017402411aba577634d8e1497363}, {pid=37, ppid=35, state=RUNNABLE; SnapshotRegionProcedure ef5fe42efd5a386cbc2993c0042cbfb9}] 2025-01-08T04:59:04,242 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=37, ppid=35, state=RUNNABLE; SnapshotRegionProcedure ef5fe42efd5a386cbc2993c0042cbfb9 2025-01-08T04:59:04,242 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=36, ppid=35, state=RUNNABLE; SnapshotRegionProcedure 87d1017402411aba577634d8e1497363 2025-01-08T04:59:04,322 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=35 2025-01-08T04:59:04,395 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T04:59:04,395 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,39801,1736312288779 2025-01-08T04:59:04,396 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=39801 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=36 2025-01-08T04:59:04,396 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37583 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=37 2025-01-08T04:59:04,396 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=36}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363. 2025-01-08T04:59:04,397 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=37}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9. 2025-01-08T04:59:04,397 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=37}] regionserver.HRegion(2538): Flush status journal for ef5fe42efd5a386cbc2993c0042cbfb9: 2025-01-08T04:59:04,397 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=36}] regionserver.HRegion(2538): Flush status journal for 87d1017402411aba577634d8e1497363: 2025-01-08T04:59:04,397 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=37}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9. for emptySnaptb0-testExportWithResetTtl completed. 2025-01-08T04:59:04,397 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=36}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363. for emptySnaptb0-testExportWithResetTtl completed. 2025-01-08T04:59:04,397 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=37}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9.' region-info for snapshot=emptySnaptb0-testExportWithResetTtl 2025-01-08T04:59:04,397 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=37}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T04:59:04,397 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=36}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363.' region-info for snapshot=emptySnaptb0-testExportWithResetTtl 2025-01-08T04:59:04,397 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=37}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T04:59:04,397 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=36}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T04:59:04,397 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=36}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T04:59:04,417 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741908_1084 (size=68) 2025-01-08T04:59:04,417 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741908_1084 (size=68) 2025-01-08T04:59:04,417 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741908_1084 (size=68) 2025-01-08T04:59:04,424 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=37}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9. 2025-01-08T04:59:04,424 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=37}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=37 2025-01-08T04:59:04,424 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=37 2025-01-08T04:59:04,424 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportWithResetTtl on region ef5fe42efd5a386cbc2993c0042cbfb9 2025-01-08T04:59:04,425 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=37, ppid=35, state=RUNNABLE; SnapshotRegionProcedure ef5fe42efd5a386cbc2993c0042cbfb9 2025-01-08T04:59:04,427 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=37, ppid=35, state=SUCCESS; SnapshotRegionProcedure ef5fe42efd5a386cbc2993c0042cbfb9 in 185 msec 2025-01-08T04:59:04,446 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741909_1085 (size=68) 2025-01-08T04:59:04,449 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741909_1085 (size=68) 2025-01-08T04:59:04,449 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741909_1085 (size=68) 2025-01-08T04:59:04,449 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=36}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363. 2025-01-08T04:59:04,449 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=36}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=36 2025-01-08T04:59:04,450 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=36 2025-01-08T04:59:04,450 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportWithResetTtl on region 87d1017402411aba577634d8e1497363 2025-01-08T04:59:04,450 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=36, ppid=35, state=RUNNABLE; SnapshotRegionProcedure 87d1017402411aba577634d8e1497363 2025-01-08T04:59:04,456 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=36, resume processing ppid=35 2025-01-08T04:59:04,456 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=36, ppid=35, state=SUCCESS; SnapshotRegionProcedure 87d1017402411aba577634d8e1497363 in 211 msec 2025-01-08T04:59:04,456 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T04:59:04,457 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T04:59:04,459 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T04:59:04,459 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T04:59:04,459 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:59:04,460 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(328): No files under family: cf 2025-01-08T04:59:04,476 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741910_1086 (size=60) 2025-01-08T04:59:04,476 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741910_1086 (size=60) 2025-01-08T04:59:04,477 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741910_1086 (size=60) 2025-01-08T04:59:04,479 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T04:59:04,479 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for emptySnaptb0-testExportWithResetTtl 2025-01-08T04:59:04,481 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/emptySnaptb0-testExportWithResetTtl 2025-01-08T04:59:04,496 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741911_1087 (size=641) 2025-01-08T04:59:04,497 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741911_1087 (size=641) 2025-01-08T04:59:04,497 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741911_1087 (size=641) 2025-01-08T04:59:04,508 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T04:59:04,525 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=35 2025-01-08T04:59:04,527 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T04:59:04,528 DEBUG [PEWorker-3 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/emptySnaptb0-testExportWithResetTtl to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/emptySnaptb0-testExportWithResetTtl 2025-01-08T04:59:04,532 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T04:59:04,532 DEBUG [PEWorker-3 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 }, snapshot procedure id = 35 2025-01-08T04:59:04,535 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=35, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } in 315 msec 2025-01-08T04:59:04,826 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=35 2025-01-08T04:59:04,827 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportWithResetTtl, procId: 35 completed 2025-01-08T04:59:04,831 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39801 {}] regionserver.HRegion(8254): writing data to region testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T04:59:04,844 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37583 {}] regionserver.HRegion(8254): writing data to region testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T04:59:04,855 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportWithResetTtl 2025-01-08T04:59:04,855 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363. 2025-01-08T04:59:04,855 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T04:59:04,915 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } 2025-01-08T04:59:04,915 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736312344915 (current time:1736312344915). 2025-01-08T04:59:04,915 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T04:59:04,915 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testExportWithResetTtl VERSION not specified, setting to 2 2025-01-08T04:59:04,915 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T04:59:04,917 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x03d73780 to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@3b96db8a 2025-01-08T04:59:04,955 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@628bf4b8, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T04:59:04,979 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T04:59:04,983 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:37896, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T04:59:04,986 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x03d73780 to 127.0.0.1:59048 2025-01-08T04:59:04,986 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T04:59:04,988 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x6fb36952 to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@3d4c0a01 2025-01-08T04:59:05,020 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7ae2356b, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T04:59:05,025 DEBUG [hconnection-0x27a33434-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T04:59:05,027 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:37912, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T04:59:05,031 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x6fb36952 to 127.0.0.1:59048 2025-01-08T04:59:05,031 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T04:59:05,031 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportWithResetTtl], kv [jenkins: RWXCA] 2025-01-08T04:59:05,032 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T04:59:05,034 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=38, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } 2025-01-08T04:59:05,035 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 }, snapshot procedure id = 38 2025-01-08T04:59:05,036 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=38 2025-01-08T04:59:05,036 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T04:59:05,038 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T04:59:05,042 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T04:59:05,064 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741912_1088 (size=156) 2025-01-08T04:59:05,064 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741912_1088 (size=156) 2025-01-08T04:59:05,064 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741912_1088 (size=156) 2025-01-08T04:59:05,070 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T04:59:05,070 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=39, ppid=38, state=RUNNABLE; SnapshotRegionProcedure 87d1017402411aba577634d8e1497363}, {pid=40, ppid=38, state=RUNNABLE; SnapshotRegionProcedure ef5fe42efd5a386cbc2993c0042cbfb9}] 2025-01-08T04:59:05,072 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=40, ppid=38, state=RUNNABLE; SnapshotRegionProcedure ef5fe42efd5a386cbc2993c0042cbfb9 2025-01-08T04:59:05,072 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=39, ppid=38, state=RUNNABLE; SnapshotRegionProcedure 87d1017402411aba577634d8e1497363 2025-01-08T04:59:05,138 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=38 2025-01-08T04:59:05,225 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,39801,1736312288779 2025-01-08T04:59:05,225 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T04:59:05,226 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=39801 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=39 2025-01-08T04:59:05,226 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37583 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=40 2025-01-08T04:59:05,226 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9. 2025-01-08T04:59:05,226 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363. 2025-01-08T04:59:05,227 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.HRegion(2837): Flushing ef5fe42efd5a386cbc2993c0042cbfb9 1/1 column families, dataSize=3.13 KB heapSize=7 KB 2025-01-08T04:59:05,227 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.HRegion(2837): Flushing 87d1017402411aba577634d8e1497363 1/1 column families, dataSize=132 B heapSize=544 B 2025-01-08T04:59:05,271 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b2025010836520c84ca874715adc07dc3d10b9fb2_ef5fe42efd5a386cbc2993c0042cbfb9 is 71, key is 11b98146a3a8f808565ac916d20e433a/cf:q/1736312344843/Put/seqid=0 2025-01-08T04:59:05,275 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e202501085d0556d645024f99a1cbc6cd1fa72745_87d1017402411aba577634d8e1497363 is 71, key is 0305c5c672d557215084c0bfeb56704c/cf:q/1736312344831/Put/seqid=0 2025-01-08T04:59:05,339 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=38 2025-01-08T04:59:05,351 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741913_1089 (size=8241) 2025-01-08T04:59:05,351 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741913_1089 (size=8241) 2025-01-08T04:59:05,351 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741913_1089 (size=8241) 2025-01-08T04:59:05,377 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741914_1090 (size=5032) 2025-01-08T04:59:05,377 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741914_1090 (size=5032) 2025-01-08T04:59:05,377 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741914_1090 (size=5032) 2025-01-08T04:59:05,380 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:59:05,465 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e202501085d0556d645024f99a1cbc6cd1fa72745_87d1017402411aba577634d8e1497363 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/d41d8cd98f00b204e9800998ecf8427e202501085d0556d645024f99a1cbc6cd1fa72745_87d1017402411aba577634d8e1497363 2025-01-08T04:59:05,474 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/87d1017402411aba577634d8e1497363/.tmp/cf/98d17a330d004eb3957f4dc65826f7a3, store: [table=testtb-testExportWithResetTtl family=cf region=87d1017402411aba577634d8e1497363] 2025-01-08T04:59:05,476 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/87d1017402411aba577634d8e1497363/.tmp/cf/98d17a330d004eb3957f4dc65826f7a3 is 206, key is 0a64d22e7bd79ff70a0e1f3c1c32297a4/cf:q/1736312344831/Put/seqid=0 2025-01-08T04:59:05,549 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741915_1091 (size=5700) 2025-01-08T04:59:05,552 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741915_1091 (size=5700) 2025-01-08T04:59:05,553 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741915_1091 (size=5700) 2025-01-08T04:59:05,564 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=132, hasBloomFilter=true, into tmp file hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/87d1017402411aba577634d8e1497363/.tmp/cf/98d17a330d004eb3957f4dc65826f7a3 2025-01-08T04:59:05,586 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/87d1017402411aba577634d8e1497363/.tmp/cf/98d17a330d004eb3957f4dc65826f7a3 as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/87d1017402411aba577634d8e1497363/cf/98d17a330d004eb3957f4dc65826f7a3 2025-01-08T04:59:05,595 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/87d1017402411aba577634d8e1497363/cf/98d17a330d004eb3957f4dc65826f7a3, entries=2, sequenceid=6, filesize=5.6 K 2025-01-08T04:59:05,596 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.HRegion(3040): Finished flush of dataSize ~132 B/132, heapSize ~528 B/528, currentSize=0 B/0 for 87d1017402411aba577634d8e1497363 in 370ms, sequenceid=6, compaction requested=false 2025-01-08T04:59:05,596 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportWithResetTtl' 2025-01-08T04:59:05,597 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.HRegion(2538): Flush status journal for 87d1017402411aba577634d8e1497363: 2025-01-08T04:59:05,597 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363. for snaptb0-testExportWithResetTtl completed. 2025-01-08T04:59:05,598 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363.' region-info for snapshot=snaptb0-testExportWithResetTtl 2025-01-08T04:59:05,598 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T04:59:05,598 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/87d1017402411aba577634d8e1497363/cf/98d17a330d004eb3957f4dc65826f7a3] hfiles 2025-01-08T04:59:05,598 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/87d1017402411aba577634d8e1497363/cf/98d17a330d004eb3957f4dc65826f7a3 for snapshot=snaptb0-testExportWithResetTtl 2025-01-08T04:59:05,612 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741916_1092 (size=107) 2025-01-08T04:59:05,613 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741916_1092 (size=107) 2025-01-08T04:59:05,613 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741916_1092 (size=107) 2025-01-08T04:59:05,614 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363. 2025-01-08T04:59:05,614 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=39 2025-01-08T04:59:05,615 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=39 2025-01-08T04:59:05,615 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportWithResetTtl on region 87d1017402411aba577634d8e1497363 2025-01-08T04:59:05,615 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=39, ppid=38, state=RUNNABLE; SnapshotRegionProcedure 87d1017402411aba577634d8e1497363 2025-01-08T04:59:05,623 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=39, ppid=38, state=SUCCESS; SnapshotRegionProcedure 87d1017402411aba577634d8e1497363 in 549 msec 2025-01-08T04:59:05,641 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=38 2025-01-08T04:59:05,759 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:59:05,772 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b2025010836520c84ca874715adc07dc3d10b9fb2_ef5fe42efd5a386cbc2993c0042cbfb9 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/c4ca4238a0b923820dcc509a6f75849b2025010836520c84ca874715adc07dc3d10b9fb2_ef5fe42efd5a386cbc2993c0042cbfb9 2025-01-08T04:59:05,774 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/ef5fe42efd5a386cbc2993c0042cbfb9/.tmp/cf/608d3bf39dcf4b74bd3a2536a19864ca, store: [table=testtb-testExportWithResetTtl family=cf region=ef5fe42efd5a386cbc2993c0042cbfb9] 2025-01-08T04:59:05,775 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/ef5fe42efd5a386cbc2993c0042cbfb9/.tmp/cf/608d3bf39dcf4b74bd3a2536a19864ca is 206, key is 1774fee741c2ac5fac5f1a60a34b283de/cf:q/1736312344843/Put/seqid=0 2025-01-08T04:59:05,787 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T04:59:05,790 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741917_1093 (size=15055) 2025-01-08T04:59:05,799 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741917_1093 (size=15055) 2025-01-08T04:59:05,799 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741917_1093 (size=15055) 2025-01-08T04:59:05,799 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=3.1 K, hasBloomFilter=true, into tmp file hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/ef5fe42efd5a386cbc2993c0042cbfb9/.tmp/cf/608d3bf39dcf4b74bd3a2536a19864ca 2025-01-08T04:59:05,808 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/ef5fe42efd5a386cbc2993c0042cbfb9/.tmp/cf/608d3bf39dcf4b74bd3a2536a19864ca as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/ef5fe42efd5a386cbc2993c0042cbfb9/cf/608d3bf39dcf4b74bd3a2536a19864ca 2025-01-08T04:59:05,826 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/ef5fe42efd5a386cbc2993c0042cbfb9/cf/608d3bf39dcf4b74bd3a2536a19864ca, entries=48, sequenceid=6, filesize=14.7 K 2025-01-08T04:59:05,829 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.HRegion(3040): Finished flush of dataSize ~3.13 KB/3204, heapSize ~6.98 KB/7152, currentSize=0 B/0 for ef5fe42efd5a386cbc2993c0042cbfb9 in 603ms, sequenceid=6, compaction requested=false 2025-01-08T04:59:05,829 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.HRegion(2538): Flush status journal for ef5fe42efd5a386cbc2993c0042cbfb9: 2025-01-08T04:59:05,829 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9. for snaptb0-testExportWithResetTtl completed. 2025-01-08T04:59:05,830 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9.' region-info for snapshot=snaptb0-testExportWithResetTtl 2025-01-08T04:59:05,830 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T04:59:05,830 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/ef5fe42efd5a386cbc2993c0042cbfb9/cf/608d3bf39dcf4b74bd3a2536a19864ca] hfiles 2025-01-08T04:59:05,830 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/ef5fe42efd5a386cbc2993c0042cbfb9/cf/608d3bf39dcf4b74bd3a2536a19864ca for snapshot=snaptb0-testExportWithResetTtl 2025-01-08T04:59:05,852 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741918_1094 (size=107) 2025-01-08T04:59:05,853 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741918_1094 (size=107) 2025-01-08T04:59:05,856 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9. 2025-01-08T04:59:05,856 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=40 2025-01-08T04:59:05,857 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=40 2025-01-08T04:59:05,857 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportWithResetTtl on region ef5fe42efd5a386cbc2993c0042cbfb9 2025-01-08T04:59:05,858 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=40, ppid=38, state=RUNNABLE; SnapshotRegionProcedure ef5fe42efd5a386cbc2993c0042cbfb9 2025-01-08T04:59:05,860 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741918_1094 (size=107) 2025-01-08T04:59:05,883 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=40, resume processing ppid=38 2025-01-08T04:59:05,883 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=40, ppid=38, state=SUCCESS; SnapshotRegionProcedure ef5fe42efd5a386cbc2993c0042cbfb9 in 793 msec 2025-01-08T04:59:05,883 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T04:59:05,884 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T04:59:05,886 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T04:59:05,886 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T04:59:05,886 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:59:05,892 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/c4ca4238a0b923820dcc509a6f75849b2025010836520c84ca874715adc07dc3d10b9fb2_ef5fe42efd5a386cbc2993c0042cbfb9, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/d41d8cd98f00b204e9800998ecf8427e202501085d0556d645024f99a1cbc6cd1fa72745_87d1017402411aba577634d8e1497363] hfiles 2025-01-08T04:59:05,892 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/c4ca4238a0b923820dcc509a6f75849b2025010836520c84ca874715adc07dc3d10b9fb2_ef5fe42efd5a386cbc2993c0042cbfb9 2025-01-08T04:59:05,892 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/d41d8cd98f00b204e9800998ecf8427e202501085d0556d645024f99a1cbc6cd1fa72745_87d1017402411aba577634d8e1497363 2025-01-08T04:59:05,934 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741919_1095 (size=291) 2025-01-08T04:59:05,935 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741919_1095 (size=291) 2025-01-08T04:59:05,935 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741919_1095 (size=291) 2025-01-08T04:59:06,143 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=38 2025-01-08T04:59:06,354 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T04:59:06,355 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testExportWithResetTtl 2025-01-08T04:59:06,359 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/snaptb0-testExportWithResetTtl 2025-01-08T04:59:06,480 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741920_1096 (size=951) 2025-01-08T04:59:06,491 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741920_1096 (size=951) 2025-01-08T04:59:06,491 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741920_1096 (size=951) 2025-01-08T04:59:06,540 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T04:59:06,569 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T04:59:06,570 DEBUG [PEWorker-3 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/snaptb0-testExportWithResetTtl to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportWithResetTtl 2025-01-08T04:59:06,573 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T04:59:06,573 DEBUG [PEWorker-3 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 }, snapshot procedure id = 38 2025-01-08T04:59:06,583 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=38, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } in 1.5410 sec 2025-01-08T04:59:06,643 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0001_000001 (auth:SIMPLE) from 127.0.0.1:56124 2025-01-08T04:59:06,646 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-1_1/usercache/jenkins/appcache/application_1736312298603_0001/container_1736312298603_0001_01_000001/launch_container.sh] 2025-01-08T04:59:06,646 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-1_1/usercache/jenkins/appcache/application_1736312298603_0001/container_1736312298603_0001_01_000001/container_tokens] 2025-01-08T04:59:06,646 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-1_1/usercache/jenkins/appcache/application_1736312298603_0001/container_1736312298603_0001_01_000001/sysfs] 2025-01-08T04:59:07,145 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=38 2025-01-08T04:59:07,145 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportWithResetTtl, procId: 38 completed 2025-01-08T04:59:07,147 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.3 create 'testExportWithResetTtl', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T04:59:07,155 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=41, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testExportWithResetTtl 2025-01-08T04:59:07,156 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=41, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testExportWithResetTtl execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T04:59:07,159 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=41, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testExportWithResetTtl execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T04:59:07,166 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "testExportWithResetTtl" procId is: 41 2025-01-08T04:59:07,167 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=41 2025-01-08T04:59:07,256 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741921_1097 (size=433) 2025-01-08T04:59:07,256 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741921_1097 (size=433) 2025-01-08T04:59:07,257 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741921_1097 (size=433) 2025-01-08T04:59:07,263 INFO [RegionOpenAndInit-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 0af06f556013410d600de6c5efac9938, NAME => 'testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testExportWithResetTtl', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T04:59:07,263 INFO [RegionOpenAndInit-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => 24dd8210f60c73eb6228c36913fa47cb, NAME => 'testExportWithResetTtl,1,1736312347147.24dd8210f60c73eb6228c36913fa47cb.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testExportWithResetTtl', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T04:59:07,268 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=41 2025-01-08T04:59:07,345 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741922_1098 (size=58) 2025-01-08T04:59:07,346 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741922_1098 (size=58) 2025-01-08T04:59:07,347 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(894): Instantiated testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T04:59:07,348 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1681): Closing 0af06f556013410d600de6c5efac9938, disabling compactions & flushes 2025-01-08T04:59:07,348 INFO [RegionOpenAndInit-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1703): Closing region testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938. 2025-01-08T04:59:07,348 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938. 2025-01-08T04:59:07,348 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938. after waiting 0 ms 2025-01-08T04:59:07,348 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938. 2025-01-08T04:59:07,348 INFO [RegionOpenAndInit-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1922): Closed testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938. 2025-01-08T04:59:07,348 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1635): Region close journal for 0af06f556013410d600de6c5efac9938: 2025-01-08T04:59:07,349 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741922_1098 (size=58) 2025-01-08T04:59:07,408 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741923_1099 (size=58) 2025-01-08T04:59:07,408 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741923_1099 (size=58) 2025-01-08T04:59:07,408 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741923_1099 (size=58) 2025-01-08T04:59:07,410 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(894): Instantiated testExportWithResetTtl,1,1736312347147.24dd8210f60c73eb6228c36913fa47cb.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T04:59:07,410 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1681): Closing 24dd8210f60c73eb6228c36913fa47cb, disabling compactions & flushes 2025-01-08T04:59:07,410 INFO [RegionOpenAndInit-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1703): Closing region testExportWithResetTtl,1,1736312347147.24dd8210f60c73eb6228c36913fa47cb. 2025-01-08T04:59:07,410 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testExportWithResetTtl,1,1736312347147.24dd8210f60c73eb6228c36913fa47cb. 2025-01-08T04:59:07,410 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testExportWithResetTtl,1,1736312347147.24dd8210f60c73eb6228c36913fa47cb. after waiting 0 ms 2025-01-08T04:59:07,410 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testExportWithResetTtl,1,1736312347147.24dd8210f60c73eb6228c36913fa47cb. 2025-01-08T04:59:07,410 INFO [RegionOpenAndInit-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1922): Closed testExportWithResetTtl,1,1736312347147.24dd8210f60c73eb6228c36913fa47cb. 2025-01-08T04:59:07,411 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1635): Region close journal for 24dd8210f60c73eb6228c36913fa47cb: 2025-01-08T04:59:07,412 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=41, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testExportWithResetTtl execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T04:59:07,413 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938.","families":{"info":[{"qualifier":"regioninfo","vlen":57,"tag":[],"timestamp":"1736312347412"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736312347412"}]},"ts":"1736312347412"} 2025-01-08T04:59:07,413 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testExportWithResetTtl,1,1736312347147.24dd8210f60c73eb6228c36913fa47cb.","families":{"info":[{"qualifier":"regioninfo","vlen":57,"tag":[],"timestamp":"1736312347412"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736312347412"}]},"ts":"1736312347412"} 2025-01-08T04:59:07,448 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T04:59:07,450 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=41, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testExportWithResetTtl execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T04:59:07,450 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312347450"}]},"ts":"1736312347450"} 2025-01-08T04:59:07,452 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=testExportWithResetTtl, state=ENABLING in hbase:meta 2025-01-08T04:59:07,458 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(202): Hosts are {f28094b72178=0} racks are {/default-rack=0} 2025-01-08T04:59:07,460 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T04:59:07,460 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T04:59:07,460 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T04:59:07,461 INFO [PEWorker-5 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T04:59:07,461 INFO [PEWorker-5 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T04:59:07,461 INFO [PEWorker-5 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T04:59:07,461 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T04:59:07,461 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=42, ppid=41, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testExportWithResetTtl, region=0af06f556013410d600de6c5efac9938, ASSIGN}, {pid=43, ppid=41, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testExportWithResetTtl, region=24dd8210f60c73eb6228c36913fa47cb, ASSIGN}] 2025-01-08T04:59:07,463 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=43, ppid=41, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testExportWithResetTtl, region=24dd8210f60c73eb6228c36913fa47cb, ASSIGN 2025-01-08T04:59:07,463 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=42, ppid=41, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testExportWithResetTtl, region=0af06f556013410d600de6c5efac9938, ASSIGN 2025-01-08T04:59:07,464 INFO [PEWorker-1 {}] assignment.TransitRegionStateProcedure(264): Starting pid=43, ppid=41, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testExportWithResetTtl, region=24dd8210f60c73eb6228c36913fa47cb, ASSIGN; state=OFFLINE, location=f28094b72178,37583,1736312288873; forceNewPlan=false, retain=false 2025-01-08T04:59:07,465 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=42, ppid=41, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testExportWithResetTtl, region=0af06f556013410d600de6c5efac9938, ASSIGN; state=OFFLINE, location=f28094b72178,37967,1736312288941; forceNewPlan=false, retain=false 2025-01-08T04:59:07,481 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=41 2025-01-08T04:59:07,615 INFO [f28094b72178:42195 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T04:59:07,615 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=42 updating hbase:meta row=0af06f556013410d600de6c5efac9938, regionState=OPENING, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T04:59:07,615 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=43 updating hbase:meta row=24dd8210f60c73eb6228c36913fa47cb, regionState=OPENING, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T04:59:07,620 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=44, ppid=42, state=RUNNABLE; OpenRegionProcedure 0af06f556013410d600de6c5efac9938, server=f28094b72178,37967,1736312288941}] 2025-01-08T04:59:07,622 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=45, ppid=43, state=RUNNABLE; OpenRegionProcedure 24dd8210f60c73eb6228c36913fa47cb, server=f28094b72178,37583,1736312288873}] 2025-01-08T04:59:07,776 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T04:59:07,776 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T04:59:07,781 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] handler.AssignRegionHandler(135): Open testExportWithResetTtl,1,1736312347147.24dd8210f60c73eb6228c36913fa47cb. 2025-01-08T04:59:07,782 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(7285): Opening region: {ENCODED => 24dd8210f60c73eb6228c36913fa47cb, NAME => 'testExportWithResetTtl,1,1736312347147.24dd8210f60c73eb6228c36913fa47cb.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T04:59:07,782 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(7999): Registered coprocessor service: region=testExportWithResetTtl,1,1736312347147.24dd8210f60c73eb6228c36913fa47cb. service=AccessControlService 2025-01-08T04:59:07,782 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T04:59:07,783 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testExportWithResetTtl 24dd8210f60c73eb6228c36913fa47cb 2025-01-08T04:59:07,783 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(894): Instantiated testExportWithResetTtl,1,1736312347147.24dd8210f60c73eb6228c36913fa47cb.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T04:59:07,783 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(7327): checking encryption for 24dd8210f60c73eb6228c36913fa47cb 2025-01-08T04:59:07,783 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(7330): checking classloading for 24dd8210f60c73eb6228c36913fa47cb 2025-01-08T04:59:07,784 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=41 2025-01-08T04:59:07,787 INFO [StoreOpener-24dd8210f60c73eb6228c36913fa47cb-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 24dd8210f60c73eb6228c36913fa47cb 2025-01-08T04:59:07,787 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] handler.AssignRegionHandler(135): Open testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938. 2025-01-08T04:59:07,787 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(7285): Opening region: {ENCODED => 0af06f556013410d600de6c5efac9938, NAME => 'testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T04:59:07,787 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(7999): Registered coprocessor service: region=testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938. service=AccessControlService 2025-01-08T04:59:07,788 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T04:59:07,788 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testExportWithResetTtl 0af06f556013410d600de6c5efac9938 2025-01-08T04:59:07,788 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(894): Instantiated testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T04:59:07,788 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(7327): checking encryption for 0af06f556013410d600de6c5efac9938 2025-01-08T04:59:07,788 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(7330): checking classloading for 0af06f556013410d600de6c5efac9938 2025-01-08T04:59:07,791 INFO [StoreOpener-24dd8210f60c73eb6228c36913fa47cb-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 24dd8210f60c73eb6228c36913fa47cb columnFamilyName cf 2025-01-08T04:59:07,796 INFO [StoreOpener-0af06f556013410d600de6c5efac9938-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 0af06f556013410d600de6c5efac9938 2025-01-08T04:59:07,796 DEBUG [StoreOpener-24dd8210f60c73eb6228c36913fa47cb-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:59:07,798 INFO [StoreOpener-24dd8210f60c73eb6228c36913fa47cb-1 {}] regionserver.HStore(327): Store=24dd8210f60c73eb6228c36913fa47cb/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T04:59:07,799 INFO [StoreOpener-0af06f556013410d600de6c5efac9938-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 0af06f556013410d600de6c5efac9938 columnFamilyName cf 2025-01-08T04:59:07,799 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/24dd8210f60c73eb6228c36913fa47cb 2025-01-08T04:59:07,800 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/24dd8210f60c73eb6228c36913fa47cb 2025-01-08T04:59:07,802 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(1085): writing seq id for 24dd8210f60c73eb6228c36913fa47cb 2025-01-08T04:59:07,803 DEBUG [StoreOpener-0af06f556013410d600de6c5efac9938-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:59:07,803 INFO [StoreOpener-0af06f556013410d600de6c5efac9938-1 {}] regionserver.HStore(327): Store=0af06f556013410d600de6c5efac9938/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T04:59:07,808 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/0af06f556013410d600de6c5efac9938 2025-01-08T04:59:07,808 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/0af06f556013410d600de6c5efac9938 2025-01-08T04:59:07,812 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(1085): writing seq id for 0af06f556013410d600de6c5efac9938 2025-01-08T04:59:07,813 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/24dd8210f60c73eb6228c36913fa47cb/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T04:59:07,813 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(1102): Opened 24dd8210f60c73eb6228c36913fa47cb; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=70717817, jitterRate=0.05377759039402008}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T04:59:07,814 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(1001): Region open journal for 24dd8210f60c73eb6228c36913fa47cb: 2025-01-08T04:59:07,818 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/0af06f556013410d600de6c5efac9938/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T04:59:07,819 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegionServer(2601): Post open deploy tasks for testExportWithResetTtl,1,1736312347147.24dd8210f60c73eb6228c36913fa47cb., pid=45, masterSystemTime=1736312347776 2025-01-08T04:59:07,819 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T04:59:07,819 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(1102): Opened 0af06f556013410d600de6c5efac9938; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=63181153, jitterRate=-0.05852745473384857}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T04:59:07,819 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(1001): Region open journal for 0af06f556013410d600de6c5efac9938: 2025-01-08T04:59:07,825 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegionServer(2628): Finished post open deploy task for testExportWithResetTtl,1,1736312347147.24dd8210f60c73eb6228c36913fa47cb. 2025-01-08T04:59:07,825 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] handler.AssignRegionHandler(164): Opened testExportWithResetTtl,1,1736312347147.24dd8210f60c73eb6228c36913fa47cb. 2025-01-08T04:59:07,826 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegionServer(2601): Post open deploy tasks for testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938., pid=44, masterSystemTime=1736312347776 2025-01-08T04:59:07,826 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=43 updating hbase:meta row=24dd8210f60c73eb6228c36913fa47cb, regionState=OPEN, openSeqNum=2, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T04:59:07,830 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegionServer(2628): Finished post open deploy task for testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938. 2025-01-08T04:59:07,830 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] handler.AssignRegionHandler(164): Opened testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938. 2025-01-08T04:59:07,831 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=42 updating hbase:meta row=0af06f556013410d600de6c5efac9938, regionState=OPEN, openSeqNum=2, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T04:59:07,835 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=45, resume processing ppid=43 2025-01-08T04:59:07,835 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=45, ppid=43, state=SUCCESS; OpenRegionProcedure 24dd8210f60c73eb6228c36913fa47cb, server=f28094b72178,37583,1736312288873 in 209 msec 2025-01-08T04:59:07,836 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=44, resume processing ppid=42 2025-01-08T04:59:07,836 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=44, ppid=42, state=SUCCESS; OpenRegionProcedure 0af06f556013410d600de6c5efac9938, server=f28094b72178,37967,1736312288941 in 214 msec 2025-01-08T04:59:07,838 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=43, ppid=41, state=SUCCESS; TransitRegionStateProcedure table=testExportWithResetTtl, region=24dd8210f60c73eb6228c36913fa47cb, ASSIGN in 374 msec 2025-01-08T04:59:07,840 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=42, resume processing ppid=41 2025-01-08T04:59:07,840 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=42, ppid=41, state=SUCCESS; TransitRegionStateProcedure table=testExportWithResetTtl, region=0af06f556013410d600de6c5efac9938, ASSIGN in 375 msec 2025-01-08T04:59:07,841 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=41, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testExportWithResetTtl execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T04:59:07,841 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312347841"}]},"ts":"1736312347841"} 2025-01-08T04:59:07,843 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testExportWithResetTtl, state=ENABLED in hbase:meta 2025-01-08T04:59:07,847 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=41, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testExportWithResetTtl execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T04:59:07,847 DEBUG [PEWorker-3 {}] access.PermissionStorage(175): Writing permission with rowKey testExportWithResetTtl jenkins: RWXCA 2025-01-08T04:59:07,850 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37583 {}] access.PermissionStorage(611): Read acl: entry[testExportWithResetTtl], kv [jenkins: RWXCA] 2025-01-08T04:59:07,852 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T04:59:07,852 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T04:59:07,852 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T04:59:07,852 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T04:59:07,855 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T04:59:07,855 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T04:59:07,855 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportWithResetTtl with data PBUF\x0A<\x0A\x07jenkins\x121\x08\x03"-\x0A!\x0A\x07default\x12\x16testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T04:59:07,855 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportWithResetTtl with data PBUF\x0A<\x0A\x07jenkins\x121\x08\x03"-\x0A!\x0A\x07default\x12\x16testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T04:59:07,856 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T04:59:07,856 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T04:59:07,856 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportWithResetTtl with data PBUF\x0A<\x0A\x07jenkins\x121\x08\x03"-\x0A!\x0A\x07default\x12\x16testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T04:59:07,856 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportWithResetTtl with data PBUF\x0A<\x0A\x07jenkins\x121\x08\x03"-\x0A!\x0A\x07default\x12\x16testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T04:59:07,857 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=41, state=SUCCESS; CreateTableProcedure table=testExportWithResetTtl in 707 msec 2025-01-08T04:59:08,286 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=41 2025-01-08T04:59:08,287 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testExportWithResetTtl, procId: 41 completed 2025-01-08T04:59:08,305 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testExportWithResetTtl 2025-01-08T04:59:08,305 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938. 2025-01-08T04:59:08,305 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T04:59:08,346 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testExportWithResetTtl 2025-01-08T04:59:08,346 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testExportWithResetTtl Metrics about Tables on a single HBase RegionServer 2025-01-08T04:59:08,348 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportWithResetTtl 2025-01-08T04:59:08,348 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportWithResetTtl Metrics about Tables on a single HBase RegionServer 2025-01-08T04:59:08,349 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportWithTargetName 2025-01-08T04:59:08,448 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37583 {}] regionserver.HRegion(8254): writing data to region testExportWithResetTtl,1,1736312347147.24dd8210f60c73eb6228c36913fa47cb. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T04:59:08,452 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37967 {}] regionserver.HRegion(8254): writing data to region testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T04:59:08,470 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testExportWithResetTtl 2025-01-08T04:59:08,470 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938. 2025-01-08T04:59:08,470 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T04:59:08,499 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } 2025-01-08T04:59:08,499 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736312348499 (current time:1736312348499). 2025-01-08T04:59:08,500 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb-testExportWithResetTtl VERSION not specified, setting to 2 2025-01-08T04:59:08,500 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T04:59:08,501 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x5d44d5d9 to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@7eb36a73 2025-01-08T04:59:08,531 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@27a49cfc, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T04:59:08,547 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T04:59:08,548 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:37928, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T04:59:08,550 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x5d44d5d9 to 127.0.0.1:59048 2025-01-08T04:59:08,551 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T04:59:08,552 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x5f0290b1 to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@3a0789d0 2025-01-08T04:59:08,589 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@39c7a110, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T04:59:08,619 DEBUG [hconnection-0x6f47194-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T04:59:08,620 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:37944, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T04:59:08,624 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x5f0290b1 to 127.0.0.1:59048 2025-01-08T04:59:08,624 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T04:59:08,625 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] access.PermissionStorage(611): Read acl: entry[testExportWithResetTtl], kv [jenkins: RWXCA] 2025-01-08T04:59:08,626 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T04:59:08,628 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=46, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } 2025-01-08T04:59:08,628 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 }, snapshot procedure id = 46 2025-01-08T04:59:08,629 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_PREPARE 2025-01-08T04:59:08,629 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=46 2025-01-08T04:59:08,630 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T04:59:08,641 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T04:59:08,734 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=46 2025-01-08T04:59:08,756 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741924_1100 (size=143) 2025-01-08T04:59:08,758 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741924_1100 (size=143) 2025-01-08T04:59:08,758 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741924_1100 (size=143) 2025-01-08T04:59:08,939 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=46 2025-01-08T04:59:09,163 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T04:59:09,163 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=47, ppid=46, state=RUNNABLE; SnapshotRegionProcedure 0af06f556013410d600de6c5efac9938}, {pid=48, ppid=46, state=RUNNABLE; SnapshotRegionProcedure 24dd8210f60c73eb6228c36913fa47cb}] 2025-01-08T04:59:09,165 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=48, ppid=46, state=RUNNABLE; SnapshotRegionProcedure 24dd8210f60c73eb6228c36913fa47cb 2025-01-08T04:59:09,165 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=47, ppid=46, state=RUNNABLE; SnapshotRegionProcedure 0af06f556013410d600de6c5efac9938 2025-01-08T04:59:09,243 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=46 2025-01-08T04:59:09,317 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T04:59:09,319 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=37583 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=48 2025-01-08T04:59:09,320 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testExportWithResetTtl,1,1736312347147.24dd8210f60c73eb6228c36913fa47cb. 2025-01-08T04:59:09,320 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.HRegion(2837): Flushing 24dd8210f60c73eb6228c36913fa47cb 1/1 column families, dataSize=3.00 KB heapSize=6.72 KB 2025-01-08T04:59:09,321 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T04:59:09,321 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37967 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=47 2025-01-08T04:59:09,331 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938. 2025-01-08T04:59:09,331 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.HRegion(2837): Flushing 0af06f556013410d600de6c5efac9938 1/1 column families, dataSize=266 B heapSize=832 B 2025-01-08T04:59:09,382 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108f6cd2acff11b4af086908557b3100c41_24dd8210f60c73eb6228c36913fa47cb is 71, key is 108c3967a7471e6370b7e5240903d3fa/cf:q/1736312348448/Put/seqid=0 2025-01-08T04:59:09,394 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108f141ee0ac9404649bbed3b353dc79d3d_0af06f556013410d600de6c5efac9938 is 71, key is 03f5fb7191f136770c914e12750a8fcb/cf:q/1736312348452/Put/seqid=0 2025-01-08T04:59:09,492 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741925_1101 (size=5172) 2025-01-08T04:59:09,493 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741925_1101 (size=5172) 2025-01-08T04:59:09,493 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741925_1101 (size=5172) 2025-01-08T04:59:09,494 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:59:09,506 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108f141ee0ac9404649bbed3b353dc79d3d_0af06f556013410d600de6c5efac9938 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/d41d8cd98f00b204e9800998ecf8427e20250108f141ee0ac9404649bbed3b353dc79d3d_0af06f556013410d600de6c5efac9938 2025-01-08T04:59:09,511 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/0af06f556013410d600de6c5efac9938/.tmp/cf/5c9fe64a27ac422caea23e89b1e7192b, store: [table=testExportWithResetTtl family=cf region=0af06f556013410d600de6c5efac9938] 2025-01-08T04:59:09,512 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/0af06f556013410d600de6c5efac9938/.tmp/cf/5c9fe64a27ac422caea23e89b1e7192b is 199, key is 0c6ef7972a5f8403b5be56f58a90aa867/cf:q/1736312348452/Put/seqid=0 2025-01-08T04:59:09,551 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741926_1102 (size=8101) 2025-01-08T04:59:09,555 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741926_1102 (size=8101) 2025-01-08T04:59:09,555 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741926_1102 (size=8101) 2025-01-08T04:59:09,564 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:59:09,594 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108f6cd2acff11b4af086908557b3100c41_24dd8210f60c73eb6228c36913fa47cb to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/c4ca4238a0b923820dcc509a6f75849b20250108f6cd2acff11b4af086908557b3100c41_24dd8210f60c73eb6228c36913fa47cb 2025-01-08T04:59:09,599 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/24dd8210f60c73eb6228c36913fa47cb/.tmp/cf/0d20f1a10eff40fd91509e2598c3331c, store: [table=testExportWithResetTtl family=cf region=24dd8210f60c73eb6228c36913fa47cb] 2025-01-08T04:59:09,600 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/24dd8210f60c73eb6228c36913fa47cb/.tmp/cf/0d20f1a10eff40fd91509e2598c3331c is 199, key is 1d5bdc98a42cacbb4df7aa68db786d6de/cf:q/1736312348448/Put/seqid=0 2025-01-08T04:59:09,612 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741927_1103 (size=6073) 2025-01-08T04:59:09,612 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741927_1103 (size=6073) 2025-01-08T04:59:09,615 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741927_1103 (size=6073) 2025-01-08T04:59:09,619 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=5, memsize=266, hasBloomFilter=true, into tmp file hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/0af06f556013410d600de6c5efac9938/.tmp/cf/5c9fe64a27ac422caea23e89b1e7192b 2025-01-08T04:59:09,646 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/0af06f556013410d600de6c5efac9938/.tmp/cf/5c9fe64a27ac422caea23e89b1e7192b as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/0af06f556013410d600de6c5efac9938/cf/5c9fe64a27ac422caea23e89b1e7192b 2025-01-08T04:59:09,661 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741928_1104 (size=14322) 2025-01-08T04:59:09,665 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741928_1104 (size=14322) 2025-01-08T04:59:09,665 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741928_1104 (size=14322) 2025-01-08T04:59:09,671 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=5, memsize=3.0 K, hasBloomFilter=true, into tmp file hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/24dd8210f60c73eb6228c36913fa47cb/.tmp/cf/0d20f1a10eff40fd91509e2598c3331c 2025-01-08T04:59:09,675 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/0af06f556013410d600de6c5efac9938/cf/5c9fe64a27ac422caea23e89b1e7192b, entries=4, sequenceid=5, filesize=5.9 K 2025-01-08T04:59:09,681 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.HRegion(3040): Finished flush of dataSize ~266 B/266, heapSize ~816 B/816, currentSize=0 B/0 for 0af06f556013410d600de6c5efac9938 in 350ms, sequenceid=5, compaction requested=false 2025-01-08T04:59:09,682 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testExportWithResetTtl' 2025-01-08T04:59:09,682 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.HRegion(2538): Flush status journal for 0af06f556013410d600de6c5efac9938: 2025-01-08T04:59:09,682 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.SnapshotRegionCallable(75): Snapshotting region testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938. for snaptb-testExportWithResetTtl completed. 2025-01-08T04:59:09,683 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] snapshot.SnapshotManifest(241): Storing 'testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938.' region-info for snapshot=snaptb-testExportWithResetTtl 2025-01-08T04:59:09,683 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T04:59:09,683 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/0af06f556013410d600de6c5efac9938/cf/5c9fe64a27ac422caea23e89b1e7192b] hfiles 2025-01-08T04:59:09,683 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/0af06f556013410d600de6c5efac9938/cf/5c9fe64a27ac422caea23e89b1e7192b for snapshot=snaptb-testExportWithResetTtl 2025-01-08T04:59:09,703 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/24dd8210f60c73eb6228c36913fa47cb/.tmp/cf/0d20f1a10eff40fd91509e2598c3331c as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/24dd8210f60c73eb6228c36913fa47cb/cf/0d20f1a10eff40fd91509e2598c3331c 2025-01-08T04:59:09,746 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=46 2025-01-08T04:59:09,773 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/24dd8210f60c73eb6228c36913fa47cb/cf/0d20f1a10eff40fd91509e2598c3331c, entries=46, sequenceid=5, filesize=14.0 K 2025-01-08T04:59:09,783 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.HRegion(3040): Finished flush of dataSize ~3.00 KB/3070, heapSize ~6.70 KB/6864, currentSize=0 B/0 for 24dd8210f60c73eb6228c36913fa47cb in 463ms, sequenceid=5, compaction requested=false 2025-01-08T04:59:09,783 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.HRegion(2538): Flush status journal for 24dd8210f60c73eb6228c36913fa47cb: 2025-01-08T04:59:09,783 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.SnapshotRegionCallable(75): Snapshotting region testExportWithResetTtl,1,1736312347147.24dd8210f60c73eb6228c36913fa47cb. for snaptb-testExportWithResetTtl completed. 2025-01-08T04:59:09,783 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] snapshot.SnapshotManifest(241): Storing 'testExportWithResetTtl,1,1736312347147.24dd8210f60c73eb6228c36913fa47cb.' region-info for snapshot=snaptb-testExportWithResetTtl 2025-01-08T04:59:09,783 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T04:59:09,783 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/24dd8210f60c73eb6228c36913fa47cb/cf/0d20f1a10eff40fd91509e2598c3331c] hfiles 2025-01-08T04:59:09,783 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/24dd8210f60c73eb6228c36913fa47cb/cf/0d20f1a10eff40fd91509e2598c3331c for snapshot=snaptb-testExportWithResetTtl 2025-01-08T04:59:09,814 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741929_1105 (size=100) 2025-01-08T04:59:09,814 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741929_1105 (size=100) 2025-01-08T04:59:09,814 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741929_1105 (size=100) 2025-01-08T04:59:09,900 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741930_1106 (size=100) 2025-01-08T04:59:09,900 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741930_1106 (size=100) 2025-01-08T04:59:09,900 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741930_1106 (size=100) 2025-01-08T04:59:09,906 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testExportWithResetTtl,1,1736312347147.24dd8210f60c73eb6228c36913fa47cb. 2025-01-08T04:59:09,907 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=48 2025-01-08T04:59:09,915 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=48 2025-01-08T04:59:09,915 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb-testExportWithResetTtl on region 24dd8210f60c73eb6228c36913fa47cb 2025-01-08T04:59:09,915 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=48, ppid=46, state=RUNNABLE; SnapshotRegionProcedure 24dd8210f60c73eb6228c36913fa47cb 2025-01-08T04:59:09,943 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=48, ppid=46, state=SUCCESS; SnapshotRegionProcedure 24dd8210f60c73eb6228c36913fa47cb in 769 msec 2025-01-08T04:59:10,218 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938. 2025-01-08T04:59:10,218 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=47 2025-01-08T04:59:10,219 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=47 2025-01-08T04:59:10,219 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb-testExportWithResetTtl on region 0af06f556013410d600de6c5efac9938 2025-01-08T04:59:10,219 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=47, ppid=46, state=RUNNABLE; SnapshotRegionProcedure 0af06f556013410d600de6c5efac9938 2025-01-08T04:59:10,224 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=47, resume processing ppid=46 2025-01-08T04:59:10,224 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T04:59:10,224 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=47, ppid=46, state=SUCCESS; SnapshotRegionProcedure 0af06f556013410d600de6c5efac9938 in 1.0580 sec 2025-01-08T04:59:10,225 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T04:59:10,226 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T04:59:10,227 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T04:59:10,227 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:59:10,228 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/c4ca4238a0b923820dcc509a6f75849b20250108f6cd2acff11b4af086908557b3100c41_24dd8210f60c73eb6228c36913fa47cb, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/d41d8cd98f00b204e9800998ecf8427e20250108f141ee0ac9404649bbed3b353dc79d3d_0af06f556013410d600de6c5efac9938] hfiles 2025-01-08T04:59:10,229 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/c4ca4238a0b923820dcc509a6f75849b20250108f6cd2acff11b4af086908557b3100c41_24dd8210f60c73eb6228c36913fa47cb 2025-01-08T04:59:10,229 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/d41d8cd98f00b204e9800998ecf8427e20250108f141ee0ac9404649bbed3b353dc79d3d_0af06f556013410d600de6c5efac9938 2025-01-08T04:59:10,281 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741931_1107 (size=284) 2025-01-08T04:59:10,281 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741931_1107 (size=284) 2025-01-08T04:59:10,281 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741931_1107 (size=284) 2025-01-08T04:59:10,293 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T04:59:10,293 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb-testExportWithResetTtl 2025-01-08T04:59:10,295 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/snaptb-testExportWithResetTtl 2025-01-08T04:59:10,322 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741932_1108 (size=923) 2025-01-08T04:59:10,323 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741932_1108 (size=923) 2025-01-08T04:59:10,323 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741932_1108 (size=923) 2025-01-08T04:59:10,333 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T04:59:10,344 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T04:59:10,344 DEBUG [PEWorker-1 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/snaptb-testExportWithResetTtl to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb-testExportWithResetTtl 2025-01-08T04:59:10,346 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T04:59:10,346 DEBUG [PEWorker-1 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 }, snapshot procedure id = 46 2025-01-08T04:59:10,348 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=46, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } in 1.7200 sec 2025-01-08T04:59:10,755 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=46 2025-01-08T04:59:10,755 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testExportWithResetTtl, procId: 46 completed 2025-01-08T04:59:10,783 INFO [Time-limited test {}] snapshot.TestExportSnapshot(468): HDFS export destination path: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312350783 2025-01-08T04:59:10,783 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=hdfs://localhost:45939, tgtDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312350783, rawTgtDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312350783, srcFsUri=hdfs://localhost:45939, srcDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T04:59:10,869 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:45939, inputRoot=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T04:59:10,869 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1149249371_22, ugi=jenkins (auth:SIMPLE)]], outputRoot=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312350783, skipTmp=false, initialOutputSnapshotDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312350783/.hbase-snapshot/.tmp/snaptb-testExportWithResetTtl 2025-01-08T04:59:10,873 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T04:59:10,926 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb-testExportWithResetTtl to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312350783/.hbase-snapshot/.tmp/snaptb-testExportWithResetTtl 2025-01-08T04:59:11,081 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741934_1110 (size=143) 2025-01-08T04:59:11,081 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741934_1110 (size=143) 2025-01-08T04:59:11,081 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741934_1110 (size=143) 2025-01-08T04:59:11,083 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741933_1109 (size=923) 2025-01-08T04:59:11,090 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741933_1109 (size=923) 2025-01-08T04:59:11,090 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741933_1109 (size=923) 2025-01-08T04:59:11,557 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741935_1111 (size=141) 2025-01-08T04:59:11,558 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741935_1111 (size=141) 2025-01-08T04:59:11,558 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741935_1111 (size=141) 2025-01-08T04:59:11,561 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.7.0-SNAPSHOT.jar 2025-01-08T04:59:11,562 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T04:59:11,562 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T04:59:11,562 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T04:59:13,142 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop-1624811534777123645.jar 2025-01-08T04:59:13,143 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T04:59:13,144 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T04:59:13,264 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop-16649232656014816322.jar 2025-01-08T04:59:13,265 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T04:59:13,265 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T04:59:13,265 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T04:59:13,266 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T04:59:13,266 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T04:59:13,267 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T04:59:13,267 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T04:59:13,267 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T04:59:13,268 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T04:59:13,268 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T04:59:13,268 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T04:59:13,269 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T04:59:13,269 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T04:59:13,269 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T04:59:13,270 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T04:59:13,270 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T04:59:13,270 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T04:59:13,271 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T04:59:13,271 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T04:59:13,272 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T04:59:13,272 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T04:59:13,272 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T04:59:13,273 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T04:59:13,273 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T04:59:13,273 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T04:59:13,287 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T04:59:13,480 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741936_1112 (size=127628) 2025-01-08T04:59:13,480 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741936_1112 (size=127628) 2025-01-08T04:59:13,487 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741936_1112 (size=127628) 2025-01-08T04:59:13,568 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741937_1113 (size=2172137) 2025-01-08T04:59:13,568 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741937_1113 (size=2172137) 2025-01-08T04:59:13,569 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741937_1113 (size=2172137) 2025-01-08T04:59:13,660 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741938_1114 (size=213228) 2025-01-08T04:59:13,663 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741938_1114 (size=213228) 2025-01-08T04:59:13,664 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741938_1114 (size=213228) 2025-01-08T04:59:13,798 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741939_1115 (size=1877034) 2025-01-08T04:59:13,799 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741939_1115 (size=1877034) 2025-01-08T04:59:13,803 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741939_1115 (size=1877034) 2025-01-08T04:59:13,886 DEBUG [master/f28094b72178:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region c1df98ea008addbfc51d7e8682ea8fd9 changed from -1.0 to 0.0, refreshing cache 2025-01-08T04:59:13,886 DEBUG [master/f28094b72178:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 52312176a1351392d55e81604512c13b changed from -1.0 to 0.0, refreshing cache 2025-01-08T04:59:13,886 DEBUG [master/f28094b72178:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 24dd8210f60c73eb6228c36913fa47cb changed from -1.0 to 0.0, refreshing cache 2025-01-08T04:59:13,887 DEBUG [master/f28094b72178:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region ef5fe42efd5a386cbc2993c0042cbfb9 changed from -1.0 to 0.0, refreshing cache 2025-01-08T04:59:13,887 DEBUG [master/f28094b72178:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 0af06f556013410d600de6c5efac9938 changed from -1.0 to 0.0, refreshing cache 2025-01-08T04:59:13,887 DEBUG [master/f28094b72178:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 87d1017402411aba577634d8e1497363 changed from -1.0 to 0.0, refreshing cache 2025-01-08T04:59:14,267 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741940_1116 (size=533455) 2025-01-08T04:59:14,267 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741940_1116 (size=533455) 2025-01-08T04:59:14,270 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741940_1116 (size=533455) 2025-01-08T04:59:14,470 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741941_1117 (size=7280644) 2025-01-08T04:59:14,472 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741941_1117 (size=7280644) 2025-01-08T04:59:14,472 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741941_1117 (size=7280644) 2025-01-08T04:59:14,587 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741942_1118 (size=4188619) 2025-01-08T04:59:14,591 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741942_1118 (size=4188619) 2025-01-08T04:59:14,593 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741942_1118 (size=4188619) 2025-01-08T04:59:14,657 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741943_1119 (size=20406) 2025-01-08T04:59:14,658 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741943_1119 (size=20406) 2025-01-08T04:59:14,658 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741943_1119 (size=20406) 2025-01-08T04:59:14,711 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741944_1120 (size=75495) 2025-01-08T04:59:14,712 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741944_1120 (size=75495) 2025-01-08T04:59:14,713 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741944_1120 (size=75495) 2025-01-08T04:59:14,801 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741945_1121 (size=45609) 2025-01-08T04:59:14,802 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741945_1121 (size=45609) 2025-01-08T04:59:14,802 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741945_1121 (size=45609) 2025-01-08T04:59:14,838 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741946_1122 (size=110084) 2025-01-08T04:59:14,839 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741946_1122 (size=110084) 2025-01-08T04:59:14,840 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741946_1122 (size=110084) 2025-01-08T04:59:14,870 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741947_1123 (size=1323991) 2025-01-08T04:59:14,871 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741947_1123 (size=1323991) 2025-01-08T04:59:14,871 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741947_1123 (size=1323991) 2025-01-08T04:59:15,363 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741948_1124 (size=23076) 2025-01-08T04:59:15,364 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741948_1124 (size=23076) 2025-01-08T04:59:15,364 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741948_1124 (size=23076) 2025-01-08T04:59:15,791 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741949_1125 (size=126803) 2025-01-08T04:59:15,792 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741949_1125 (size=126803) 2025-01-08T04:59:15,792 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741949_1125 (size=126803) 2025-01-08T04:59:16,237 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741950_1126 (size=322274) 2025-01-08T04:59:16,238 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741950_1126 (size=322274) 2025-01-08T04:59:16,238 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741950_1126 (size=322274) 2025-01-08T04:59:16,285 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741951_1127 (size=1832290) 2025-01-08T04:59:16,285 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741951_1127 (size=1832290) 2025-01-08T04:59:16,286 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741951_1127 (size=1832290) 2025-01-08T04:59:16,317 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741952_1128 (size=30081) 2025-01-08T04:59:16,318 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741952_1128 (size=30081) 2025-01-08T04:59:16,321 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741952_1128 (size=30081) 2025-01-08T04:59:16,369 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741953_1129 (size=53616) 2025-01-08T04:59:16,369 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741953_1129 (size=53616) 2025-01-08T04:59:16,370 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741953_1129 (size=53616) 2025-01-08T04:59:16,390 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741954_1130 (size=29229) 2025-01-08T04:59:16,391 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741954_1130 (size=29229) 2025-01-08T04:59:16,391 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741954_1130 (size=29229) 2025-01-08T04:59:16,480 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741955_1131 (size=451756) 2025-01-08T04:59:16,480 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741955_1131 (size=451756) 2025-01-08T04:59:16,483 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741955_1131 (size=451756) 2025-01-08T04:59:16,574 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741956_1132 (size=169089) 2025-01-08T04:59:16,575 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741956_1132 (size=169089) 2025-01-08T04:59:16,576 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741956_1132 (size=169089) 2025-01-08T04:59:17,055 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741957_1133 (size=5175431) 2025-01-08T04:59:17,055 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741957_1133 (size=5175431) 2025-01-08T04:59:17,056 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741957_1133 (size=5175431) 2025-01-08T04:59:17,113 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741958_1134 (size=136454) 2025-01-08T04:59:17,119 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741958_1134 (size=136454) 2025-01-08T04:59:17,120 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741958_1134 (size=136454) 2025-01-08T04:59:17,175 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741959_1135 (size=6350712) 2025-01-08T04:59:17,175 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741959_1135 (size=6350712) 2025-01-08T04:59:17,175 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741959_1135 (size=6350712) 2025-01-08T04:59:17,610 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741960_1136 (size=907470) 2025-01-08T04:59:17,611 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741960_1136 (size=907470) 2025-01-08T04:59:17,611 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741960_1136 (size=907470) 2025-01-08T04:59:17,646 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741961_1137 (size=3317408) 2025-01-08T04:59:17,646 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741961_1137 (size=3317408) 2025-01-08T04:59:17,647 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741961_1137 (size=3317408) 2025-01-08T04:59:17,686 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741962_1138 (size=503880) 2025-01-08T04:59:17,687 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741962_1138 (size=503880) 2025-01-08T04:59:17,688 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741962_1138 (size=503880) 2025-01-08T04:59:17,762 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741963_1139 (size=4695811) 2025-01-08T04:59:17,763 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741963_1139 (size=4695811) 2025-01-08T04:59:17,764 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741963_1139 (size=4695811) 2025-01-08T04:59:17,765 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T04:59:17,769 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'snaptb-testExportWithResetTtl' hfile list 2025-01-08T04:59:17,774 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(751): export split=0 size=32.9 K 2025-01-08T04:59:17,830 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741964_1140 (size=686) 2025-01-08T04:59:17,831 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741964_1140 (size=686) 2025-01-08T04:59:17,831 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741964_1140 (size=686) 2025-01-08T04:59:17,864 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741965_1141 (size=15) 2025-01-08T04:59:17,864 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741965_1141 (size=15) 2025-01-08T04:59:17,865 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741965_1141 (size=15) 2025-01-08T04:59:17,927 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741966_1142 (size=304875) 2025-01-08T04:59:17,928 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741966_1142 (size=304875) 2025-01-08T04:59:17,931 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741966_1142 (size=304875) 2025-01-08T04:59:17,968 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T04:59:17,968 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T04:59:18,376 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0002_000001 (auth:SIMPLE) from 127.0.0.1:47336 2025-01-08T04:59:28,803 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0002_000001 (auth:SIMPLE) from 127.0.0.1:47092 2025-01-08T04:59:29,158 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741967_1143 (size=350549) 2025-01-08T04:59:29,159 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741967_1143 (size=350549) 2025-01-08T04:59:29,164 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741967_1143 (size=350549) 2025-01-08T04:59:31,172 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0002_000001 (auth:SIMPLE) from 127.0.0.1:58758 2025-01-08T04:59:34,987 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741968_1144 (size=14322) 2025-01-08T04:59:34,991 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741968_1144 (size=14322) 2025-01-08T04:59:34,991 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741968_1144 (size=14322) 2025-01-08T04:59:35,087 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741969_1145 (size=8101) 2025-01-08T04:59:35,088 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741969_1145 (size=8101) 2025-01-08T04:59:35,094 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741969_1145 (size=8101) 2025-01-08T04:59:35,175 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741970_1146 (size=6073) 2025-01-08T04:59:35,175 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741970_1146 (size=6073) 2025-01-08T04:59:35,175 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741970_1146 (size=6073) 2025-01-08T04:59:35,247 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741971_1147 (size=5172) 2025-01-08T04:59:35,255 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741971_1147 (size=5172) 2025-01-08T04:59:35,255 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741971_1147 (size=5172) 2025-01-08T04:59:35,351 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741972_1148 (size=17458) 2025-01-08T04:59:35,351 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741972_1148 (size=17458) 2025-01-08T04:59:35,353 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741972_1148 (size=17458) 2025-01-08T04:59:35,431 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741973_1149 (size=461) 2025-01-08T04:59:35,432 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741973_1149 (size=461) 2025-01-08T04:59:35,432 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741973_1149 (size=461) 2025-01-08T04:59:35,517 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741974_1150 (size=17458) 2025-01-08T04:59:35,517 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741974_1150 (size=17458) 2025-01-08T04:59:35,518 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741974_1150 (size=17458) 2025-01-08T04:59:35,528 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_3/usercache/jenkins/appcache/application_1736312298603_0002/container_1736312298603_0002_01_000002/launch_container.sh] 2025-01-08T04:59:35,529 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_3/usercache/jenkins/appcache/application_1736312298603_0002/container_1736312298603_0002_01_000002/container_tokens] 2025-01-08T04:59:35,529 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_3/usercache/jenkins/appcache/application_1736312298603_0002/container_1736312298603_0002_01_000002/sysfs] 2025-01-08T04:59:35,562 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741975_1151 (size=350549) 2025-01-08T04:59:35,563 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741975_1151 (size=350549) 2025-01-08T04:59:35,564 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741975_1151 (size=350549) 2025-01-08T04:59:35,588 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0002_000001 (auth:SIMPLE) from 127.0.0.1:37970 2025-01-08T04:59:35,787 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T04:59:37,462 INFO [Time-limited test {}] snapshot.ExportSnapshot(1207): Finalize the Snapshot Export 2025-01-08T04:59:37,464 INFO [Time-limited test {}] snapshot.ExportSnapshot(1218): Verify the exported snapshot's expiration status and integrity. 2025-01-08T04:59:37,483 INFO [Time-limited test {}] snapshot.ExportSnapshot(1224): Export Completed: snaptb-testExportWithResetTtl 2025-01-08T04:59:37,483 INFO [Time-limited test {}] snapshot.TestExportSnapshot(362): Exported snapshot 2025-01-08T04:59:37,484 INFO [Time-limited test {}] snapshot.TestExportSnapshot(373): Verified filesystem state 2025-01-08T04:59:37,484 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1149249371_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb-testExportWithResetTtl at hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb-testExportWithResetTtl 2025-01-08T04:59:37,485 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb-testExportWithResetTtl/.snapshotinfo 2025-01-08T04:59:37,485 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb-testExportWithResetTtl/data.manifest 2025-01-08T04:59:37,485 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1149249371_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312350783/.hbase-snapshot/snaptb-testExportWithResetTtl at hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312350783/.hbase-snapshot/snaptb-testExportWithResetTtl 2025-01-08T04:59:37,486 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312350783/.hbase-snapshot/snaptb-testExportWithResetTtl/.snapshotinfo 2025-01-08T04:59:37,486 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312350783/.hbase-snapshot/snaptb-testExportWithResetTtl/data.manifest 2025-01-08T04:59:37,495 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testExportWithResetTtl 2025-01-08T04:59:37,496 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.3 disable testExportWithResetTtl 2025-01-08T04:59:37,498 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=49, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testExportWithResetTtl 2025-01-08T04:59:37,501 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=49 2025-01-08T04:59:37,502 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312377501"}]},"ts":"1736312377501"} 2025-01-08T04:59:37,503 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testExportWithResetTtl, state=DISABLING in hbase:meta 2025-01-08T04:59:37,574 INFO [PEWorker-2 {}] procedure.DisableTableProcedure(284): Set testExportWithResetTtl to state=DISABLING 2025-01-08T04:59:37,599 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=50, ppid=49, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testExportWithResetTtl}] 2025-01-08T04:59:37,601 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=51, ppid=50, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testExportWithResetTtl, region=0af06f556013410d600de6c5efac9938, UNASSIGN}, {pid=52, ppid=50, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testExportWithResetTtl, region=24dd8210f60c73eb6228c36913fa47cb, UNASSIGN}] 2025-01-08T04:59:37,602 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=52, ppid=50, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testExportWithResetTtl, region=24dd8210f60c73eb6228c36913fa47cb, UNASSIGN 2025-01-08T04:59:37,602 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=51, ppid=50, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testExportWithResetTtl, region=0af06f556013410d600de6c5efac9938, UNASSIGN 2025-01-08T04:59:37,602 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=49 2025-01-08T04:59:37,603 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=52 updating hbase:meta row=24dd8210f60c73eb6228c36913fa47cb, regionState=CLOSING, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T04:59:37,604 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=51 updating hbase:meta row=0af06f556013410d600de6c5efac9938, regionState=CLOSING, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T04:59:37,604 DEBUG [PEWorker-5 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T04:59:37,604 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=53, ppid=52, state=RUNNABLE; CloseRegionProcedure 24dd8210f60c73eb6228c36913fa47cb, server=f28094b72178,37583,1736312288873}] 2025-01-08T04:59:37,605 DEBUG [PEWorker-3 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T04:59:37,605 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=54, ppid=51, state=RUNNABLE; CloseRegionProcedure 0af06f556013410d600de6c5efac9938, server=f28094b72178,37967,1736312288941}] 2025-01-08T04:59:37,756 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T04:59:37,757 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] handler.UnassignRegionHandler(124): Close 24dd8210f60c73eb6228c36913fa47cb 2025-01-08T04:59:37,757 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T04:59:37,757 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] regionserver.HRegion(1681): Closing 24dd8210f60c73eb6228c36913fa47cb, disabling compactions & flushes 2025-01-08T04:59:37,758 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] regionserver.HRegion(1703): Closing region testExportWithResetTtl,1,1736312347147.24dd8210f60c73eb6228c36913fa47cb. 2025-01-08T04:59:37,758 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] regionserver.HRegion(1724): Waiting without time limit for close lock on testExportWithResetTtl,1,1736312347147.24dd8210f60c73eb6228c36913fa47cb. 2025-01-08T04:59:37,758 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] regionserver.HRegion(1791): Acquired close lock on testExportWithResetTtl,1,1736312347147.24dd8210f60c73eb6228c36913fa47cb. after waiting 0 ms 2025-01-08T04:59:37,758 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] regionserver.HRegion(1801): Updates disabled for region testExportWithResetTtl,1,1736312347147.24dd8210f60c73eb6228c36913fa47cb. 2025-01-08T04:59:37,761 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T04:59:37,762 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] handler.UnassignRegionHandler(124): Close 0af06f556013410d600de6c5efac9938 2025-01-08T04:59:37,762 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T04:59:37,762 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] regionserver.HRegion(1681): Closing 0af06f556013410d600de6c5efac9938, disabling compactions & flushes 2025-01-08T04:59:37,762 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] regionserver.HRegion(1703): Closing region testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938. 2025-01-08T04:59:37,762 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] regionserver.HRegion(1724): Waiting without time limit for close lock on testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938. 2025-01-08T04:59:37,762 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] regionserver.HRegion(1791): Acquired close lock on testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938. after waiting 0 ms 2025-01-08T04:59:37,762 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] regionserver.HRegion(1801): Updates disabled for region testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938. 2025-01-08T04:59:37,813 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=49 2025-01-08T04:59:37,855 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/0af06f556013410d600de6c5efac9938/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2025-01-08T04:59:37,873 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/24dd8210f60c73eb6228c36913fa47cb/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2025-01-08T04:59:37,874 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T04:59:37,874 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] regionserver.HRegion(1922): Closed testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938. 2025-01-08T04:59:37,875 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] regionserver.HRegion(1635): Region close journal for 0af06f556013410d600de6c5efac9938: 2025-01-08T04:59:37,875 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T04:59:37,875 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] regionserver.HRegion(1922): Closed testExportWithResetTtl,1,1736312347147.24dd8210f60c73eb6228c36913fa47cb. 2025-01-08T04:59:37,875 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] regionserver.HRegion(1635): Region close journal for 24dd8210f60c73eb6228c36913fa47cb: 2025-01-08T04:59:37,877 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] handler.UnassignRegionHandler(170): Closed 0af06f556013410d600de6c5efac9938 2025-01-08T04:59:37,880 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] handler.UnassignRegionHandler(170): Closed 24dd8210f60c73eb6228c36913fa47cb 2025-01-08T04:59:37,880 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=51 updating hbase:meta row=0af06f556013410d600de6c5efac9938, regionState=CLOSED 2025-01-08T04:59:37,880 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=52 updating hbase:meta row=24dd8210f60c73eb6228c36913fa47cb, regionState=CLOSED 2025-01-08T04:59:37,885 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=54, resume processing ppid=51 2025-01-08T04:59:37,885 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=54, ppid=51, state=SUCCESS; CloseRegionProcedure 0af06f556013410d600de6c5efac9938, server=f28094b72178,37967,1736312288941 in 278 msec 2025-01-08T04:59:37,890 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=51, ppid=50, state=SUCCESS; TransitRegionStateProcedure table=testExportWithResetTtl, region=0af06f556013410d600de6c5efac9938, UNASSIGN in 284 msec 2025-01-08T04:59:37,891 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=53, resume processing ppid=52 2025-01-08T04:59:37,892 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=53, ppid=52, state=SUCCESS; CloseRegionProcedure 24dd8210f60c73eb6228c36913fa47cb, server=f28094b72178,37583,1736312288873 in 282 msec 2025-01-08T04:59:37,894 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=52, resume processing ppid=50 2025-01-08T04:59:37,895 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=52, ppid=50, state=SUCCESS; TransitRegionStateProcedure table=testExportWithResetTtl, region=24dd8210f60c73eb6228c36913fa47cb, UNASSIGN in 291 msec 2025-01-08T04:59:37,897 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=50, resume processing ppid=49 2025-01-08T04:59:37,897 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=50, ppid=49, state=SUCCESS; CloseTableRegionsProcedure table=testExportWithResetTtl in 296 msec 2025-01-08T04:59:37,899 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312377899"}]},"ts":"1736312377899"} 2025-01-08T04:59:37,901 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testExportWithResetTtl, state=DISABLED in hbase:meta 2025-01-08T04:59:37,944 INFO [PEWorker-3 {}] procedure.DisableTableProcedure(296): Set testExportWithResetTtl to state=DISABLED 2025-01-08T04:59:37,947 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=49, state=SUCCESS; DisableTableProcedure table=testExportWithResetTtl in 449 msec 2025-01-08T04:59:38,115 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=49 2025-01-08T04:59:38,116 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testExportWithResetTtl, procId: 49 completed 2025-01-08T04:59:38,117 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.3 delete testExportWithResetTtl 2025-01-08T04:59:38,118 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=55, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testExportWithResetTtl 2025-01-08T04:59:38,119 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=55, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testExportWithResetTtl 2025-01-08T04:59:38,119 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] access.PermissionStorage(259): Removing permissions of removed table testExportWithResetTtl 2025-01-08T04:59:38,121 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=55, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testExportWithResetTtl 2025-01-08T04:59:38,122 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37583 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testExportWithResetTtl 2025-01-08T04:59:38,138 DEBUG [HFileArchiver-3 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/0af06f556013410d600de6c5efac9938 2025-01-08T04:59:38,142 DEBUG [HFileArchiver-3 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/0af06f556013410d600de6c5efac9938/cf, FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/0af06f556013410d600de6c5efac9938/recovered.edits] 2025-01-08T04:59:38,148 DEBUG [HFileArchiver-3 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/0af06f556013410d600de6c5efac9938/cf/5c9fe64a27ac422caea23e89b1e7192b to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testExportWithResetTtl/0af06f556013410d600de6c5efac9938/cf/5c9fe64a27ac422caea23e89b1e7192b 2025-01-08T04:59:38,153 DEBUG [HFileArchiver-3 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/0af06f556013410d600de6c5efac9938/recovered.edits/8.seqid to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testExportWithResetTtl/0af06f556013410d600de6c5efac9938/recovered.edits/8.seqid 2025-01-08T04:59:38,154 DEBUG [HFileArchiver-3 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/0af06f556013410d600de6c5efac9938 2025-01-08T04:59:38,154 DEBUG [HFileArchiver-4 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/24dd8210f60c73eb6228c36913fa47cb 2025-01-08T04:59:38,160 DEBUG [HFileArchiver-4 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/24dd8210f60c73eb6228c36913fa47cb/cf, FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/24dd8210f60c73eb6228c36913fa47cb/recovered.edits] 2025-01-08T04:59:38,165 DEBUG [HFileArchiver-4 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/24dd8210f60c73eb6228c36913fa47cb/cf/0d20f1a10eff40fd91509e2598c3331c to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testExportWithResetTtl/24dd8210f60c73eb6228c36913fa47cb/cf/0d20f1a10eff40fd91509e2598c3331c 2025-01-08T04:59:38,169 DEBUG [HFileArchiver-4 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/24dd8210f60c73eb6228c36913fa47cb/recovered.edits/8.seqid to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testExportWithResetTtl/24dd8210f60c73eb6228c36913fa47cb/recovered.edits/8.seqid 2025-01-08T04:59:38,170 DEBUG [HFileArchiver-4 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportWithResetTtl/24dd8210f60c73eb6228c36913fa47cb 2025-01-08T04:59:38,170 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(313): Archived testExportWithResetTtl regions 2025-01-08T04:59:38,171 DEBUG [PEWorker-4 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d 2025-01-08T04:59:38,172 DEBUG [PEWorker-4 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf] 2025-01-08T04:59:38,183 DEBUG [PEWorker-4 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/c4ca4238a0b923820dcc509a6f75849b20250108f6cd2acff11b4af086908557b3100c41_24dd8210f60c73eb6228c36913fa47cb to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/c4ca4238a0b923820dcc509a6f75849b20250108f6cd2acff11b4af086908557b3100c41_24dd8210f60c73eb6228c36913fa47cb 2025-01-08T04:59:38,186 DEBUG [PEWorker-4 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/d41d8cd98f00b204e9800998ecf8427e20250108f141ee0ac9404649bbed3b353dc79d3d_0af06f556013410d600de6c5efac9938 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/d41d8cd98f00b204e9800998ecf8427e20250108f141ee0ac9404649bbed3b353dc79d3d_0af06f556013410d600de6c5efac9938 2025-01-08T04:59:38,187 DEBUG [PEWorker-4 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d 2025-01-08T04:59:38,202 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=55, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testExportWithResetTtl 2025-01-08T04:59:38,213 WARN [PEWorker-4 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testExportWithResetTtl from hbase:meta 2025-01-08T04:59:38,290 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testExportWithResetTtl 2025-01-08T04:59:38,290 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testExportWithResetTtl 2025-01-08T04:59:38,290 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testExportWithResetTtl 2025-01-08T04:59:38,290 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testExportWithResetTtl 2025-01-08T04:59:38,291 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportWithResetTtl with data PBUF 2025-01-08T04:59:38,291 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportWithResetTtl with data PBUF 2025-01-08T04:59:38,291 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportWithResetTtl with data PBUF 2025-01-08T04:59:38,291 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportWithResetTtl with data PBUF 2025-01-08T04:59:38,292 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(407): Removing 'testExportWithResetTtl' descriptor. 2025-01-08T04:59:38,295 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=55, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testExportWithResetTtl 2025-01-08T04:59:38,295 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(397): Removing 'testExportWithResetTtl' from region states. 2025-01-08T04:59:38,295 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736312378295"}]},"ts":"9223372036854775807"} 2025-01-08T04:59:38,295 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testExportWithResetTtl,1,1736312347147.24dd8210f60c73eb6228c36913fa47cb.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736312378295"}]},"ts":"9223372036854775807"} 2025-01-08T04:59:38,307 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T04:59:38,307 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => 0af06f556013410d600de6c5efac9938, NAME => 'testExportWithResetTtl,,1736312347147.0af06f556013410d600de6c5efac9938.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => 24dd8210f60c73eb6228c36913fa47cb, NAME => 'testExportWithResetTtl,1,1736312347147.24dd8210f60c73eb6228c36913fa47cb.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T04:59:38,308 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(401): Marking 'testExportWithResetTtl' as deleted. 2025-01-08T04:59:38,311 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736312378308"}]},"ts":"9223372036854775807"} 2025-01-08T04:59:38,314 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1678): Deleted table testExportWithResetTtl state from META 2025-01-08T04:59:38,336 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testExportWithResetTtl 2025-01-08T04:59:38,336 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testExportWithResetTtl 2025-01-08T04:59:38,336 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testExportWithResetTtl 2025-01-08T04:59:38,336 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T04:59:38,336 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T04:59:38,336 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T04:59:38,336 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testExportWithResetTtl 2025-01-08T04:59:38,336 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T04:59:38,338 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=55 2025-01-08T04:59:38,345 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testExportWithResetTtl 2025-01-08T04:59:38,375 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T04:59:38,376 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T04:59:38,376 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T04:59:38,376 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T04:59:38,378 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(133): Finished pid=55, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testExportWithResetTtl 2025-01-08T04:59:38,383 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=55, state=SUCCESS; DeleteTableProcedure table=testExportWithResetTtl in 261 msec 2025-01-08T04:59:38,440 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=55 2025-01-08T04:59:38,441 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testExportWithResetTtl, procId: 55 completed 2025-01-08T04:59:38,442 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testExportWithResetTtl 2025-01-08T04:59:38,443 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.3 disable testtb-testExportWithResetTtl 2025-01-08T04:59:38,450 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=56, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testExportWithResetTtl 2025-01-08T04:59:38,454 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=56 2025-01-08T04:59:38,454 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312378454"}]},"ts":"1736312378454"} 2025-01-08T04:59:38,456 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithResetTtl, state=DISABLING in hbase:meta 2025-01-08T04:59:38,499 INFO [PEWorker-1 {}] procedure.DisableTableProcedure(284): Set testtb-testExportWithResetTtl to state=DISABLING 2025-01-08T04:59:38,502 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=57, ppid=56, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testExportWithResetTtl}] 2025-01-08T04:59:38,511 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=58, ppid=57, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=87d1017402411aba577634d8e1497363, UNASSIGN}, {pid=59, ppid=57, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=ef5fe42efd5a386cbc2993c0042cbfb9, UNASSIGN}] 2025-01-08T04:59:38,517 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=59, ppid=57, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=ef5fe42efd5a386cbc2993c0042cbfb9, UNASSIGN 2025-01-08T04:59:38,517 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=58, ppid=57, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=87d1017402411aba577634d8e1497363, UNASSIGN 2025-01-08T04:59:38,527 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=59 updating hbase:meta row=ef5fe42efd5a386cbc2993c0042cbfb9, regionState=CLOSING, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T04:59:38,528 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=58 updating hbase:meta row=87d1017402411aba577634d8e1497363, regionState=CLOSING, regionLocation=f28094b72178,39801,1736312288779 2025-01-08T04:59:38,533 DEBUG [PEWorker-2 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T04:59:38,533 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=60, ppid=59, state=RUNNABLE; CloseRegionProcedure ef5fe42efd5a386cbc2993c0042cbfb9, server=f28094b72178,37583,1736312288873}] 2025-01-08T04:59:38,534 DEBUG [PEWorker-3 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T04:59:38,535 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=61, ppid=58, state=RUNNABLE; CloseRegionProcedure 87d1017402411aba577634d8e1497363, server=f28094b72178,39801,1736312288779}] 2025-01-08T04:59:38,555 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=56 2025-01-08T04:59:38,689 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T04:59:38,690 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] handler.UnassignRegionHandler(124): Close ef5fe42efd5a386cbc2993c0042cbfb9 2025-01-08T04:59:38,690 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T04:59:38,690 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] regionserver.HRegion(1681): Closing ef5fe42efd5a386cbc2993c0042cbfb9, disabling compactions & flushes 2025-01-08T04:59:38,690 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] regionserver.HRegion(1703): Closing region testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9. 2025-01-08T04:59:38,690 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9. 2025-01-08T04:59:38,690 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9. after waiting 0 ms 2025-01-08T04:59:38,690 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9. 2025-01-08T04:59:38,693 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,39801,1736312288779 2025-01-08T04:59:38,694 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] handler.UnassignRegionHandler(124): Close 87d1017402411aba577634d8e1497363 2025-01-08T04:59:38,694 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T04:59:38,694 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] regionserver.HRegion(1681): Closing 87d1017402411aba577634d8e1497363, disabling compactions & flushes 2025-01-08T04:59:38,694 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] regionserver.HRegion(1703): Closing region testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363. 2025-01-08T04:59:38,694 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363. 2025-01-08T04:59:38,694 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363. after waiting 0 ms 2025-01-08T04:59:38,694 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363. 2025-01-08T04:59:38,700 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/87d1017402411aba577634d8e1497363/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T04:59:38,701 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T04:59:38,701 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] regionserver.HRegion(1922): Closed testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363. 2025-01-08T04:59:38,701 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] regionserver.HRegion(1635): Region close journal for 87d1017402411aba577634d8e1497363: 2025-01-08T04:59:38,714 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] handler.UnassignRegionHandler(170): Closed 87d1017402411aba577634d8e1497363 2025-01-08T04:59:38,715 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=58 updating hbase:meta row=87d1017402411aba577634d8e1497363, regionState=CLOSED 2025-01-08T04:59:38,717 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/ef5fe42efd5a386cbc2993c0042cbfb9/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T04:59:38,718 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T04:59:38,718 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] regionserver.HRegion(1922): Closed testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9. 2025-01-08T04:59:38,718 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] regionserver.HRegion(1635): Region close journal for ef5fe42efd5a386cbc2993c0042cbfb9: 2025-01-08T04:59:38,720 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=61, resume processing ppid=58 2025-01-08T04:59:38,720 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] handler.UnassignRegionHandler(170): Closed ef5fe42efd5a386cbc2993c0042cbfb9 2025-01-08T04:59:38,720 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=61, ppid=58, state=SUCCESS; CloseRegionProcedure 87d1017402411aba577634d8e1497363, server=f28094b72178,39801,1736312288779 in 183 msec 2025-01-08T04:59:38,721 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=59 updating hbase:meta row=ef5fe42efd5a386cbc2993c0042cbfb9, regionState=CLOSED 2025-01-08T04:59:38,722 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=58, ppid=57, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=87d1017402411aba577634d8e1497363, UNASSIGN in 210 msec 2025-01-08T04:59:38,725 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=60, resume processing ppid=59 2025-01-08T04:59:38,725 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=60, ppid=59, state=SUCCESS; CloseRegionProcedure ef5fe42efd5a386cbc2993c0042cbfb9, server=f28094b72178,37583,1736312288873 in 190 msec 2025-01-08T04:59:38,728 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=59, resume processing ppid=57 2025-01-08T04:59:38,728 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=59, ppid=57, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=ef5fe42efd5a386cbc2993c0042cbfb9, UNASSIGN in 215 msec 2025-01-08T04:59:38,731 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=57, resume processing ppid=56 2025-01-08T04:59:38,731 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=57, ppid=56, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testExportWithResetTtl in 227 msec 2025-01-08T04:59:38,733 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312378732"}]},"ts":"1736312378732"} 2025-01-08T04:59:38,734 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithResetTtl, state=DISABLED in hbase:meta 2025-01-08T04:59:38,743 INFO [PEWorker-5 {}] procedure.DisableTableProcedure(296): Set testtb-testExportWithResetTtl to state=DISABLED 2025-01-08T04:59:38,746 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=56, state=SUCCESS; DisableTableProcedure table=testtb-testExportWithResetTtl in 301 msec 2025-01-08T04:59:38,756 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=56 2025-01-08T04:59:38,757 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testExportWithResetTtl, procId: 56 completed 2025-01-08T04:59:38,757 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.3 delete testtb-testExportWithResetTtl 2025-01-08T04:59:38,758 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=62, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testExportWithResetTtl 2025-01-08T04:59:38,759 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=62, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportWithResetTtl 2025-01-08T04:59:38,760 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testExportWithResetTtl 2025-01-08T04:59:38,760 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=62, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testExportWithResetTtl 2025-01-08T04:59:38,762 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37583 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testExportWithResetTtl 2025-01-08T04:59:38,764 DEBUG [HFileArchiver-5 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/87d1017402411aba577634d8e1497363 2025-01-08T04:59:38,764 DEBUG [HFileArchiver-6 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/ef5fe42efd5a386cbc2993c0042cbfb9 2025-01-08T04:59:38,767 DEBUG [HFileArchiver-5 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/87d1017402411aba577634d8e1497363/cf, FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/87d1017402411aba577634d8e1497363/recovered.edits] 2025-01-08T04:59:38,767 DEBUG [HFileArchiver-6 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/ef5fe42efd5a386cbc2993c0042cbfb9/cf, FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/ef5fe42efd5a386cbc2993c0042cbfb9/recovered.edits] 2025-01-08T04:59:38,773 DEBUG [HFileArchiver-5 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/87d1017402411aba577634d8e1497363/cf/98d17a330d004eb3957f4dc65826f7a3 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportWithResetTtl/87d1017402411aba577634d8e1497363/cf/98d17a330d004eb3957f4dc65826f7a3 2025-01-08T04:59:38,773 DEBUG [HFileArchiver-6 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/ef5fe42efd5a386cbc2993c0042cbfb9/cf/608d3bf39dcf4b74bd3a2536a19864ca to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportWithResetTtl/ef5fe42efd5a386cbc2993c0042cbfb9/cf/608d3bf39dcf4b74bd3a2536a19864ca 2025-01-08T04:59:38,778 DEBUG [HFileArchiver-5 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/87d1017402411aba577634d8e1497363/recovered.edits/9.seqid to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportWithResetTtl/87d1017402411aba577634d8e1497363/recovered.edits/9.seqid 2025-01-08T04:59:38,778 DEBUG [HFileArchiver-5 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/87d1017402411aba577634d8e1497363 2025-01-08T04:59:38,778 DEBUG [HFileArchiver-6 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/ef5fe42efd5a386cbc2993c0042cbfb9/recovered.edits/9.seqid to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportWithResetTtl/ef5fe42efd5a386cbc2993c0042cbfb9/recovered.edits/9.seqid 2025-01-08T04:59:38,779 DEBUG [HFileArchiver-6 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithResetTtl/ef5fe42efd5a386cbc2993c0042cbfb9 2025-01-08T04:59:38,780 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(313): Archived testtb-testExportWithResetTtl regions 2025-01-08T04:59:38,780 DEBUG [PEWorker-3 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e 2025-01-08T04:59:38,781 DEBUG [PEWorker-3 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf] 2025-01-08T04:59:38,786 DEBUG [PEWorker-3 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/c4ca4238a0b923820dcc509a6f75849b2025010836520c84ca874715adc07dc3d10b9fb2_ef5fe42efd5a386cbc2993c0042cbfb9 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/c4ca4238a0b923820dcc509a6f75849b2025010836520c84ca874715adc07dc3d10b9fb2_ef5fe42efd5a386cbc2993c0042cbfb9 2025-01-08T04:59:38,788 DEBUG [PEWorker-3 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/d41d8cd98f00b204e9800998ecf8427e202501085d0556d645024f99a1cbc6cd1fa72745_87d1017402411aba577634d8e1497363 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/d41d8cd98f00b204e9800998ecf8427e202501085d0556d645024f99a1cbc6cd1fa72745_87d1017402411aba577634d8e1497363 2025-01-08T04:59:38,788 DEBUG [PEWorker-3 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e 2025-01-08T04:59:38,791 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=62, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testExportWithResetTtl 2025-01-08T04:59:38,795 WARN [PEWorker-3 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testtb-testExportWithResetTtl from hbase:meta 2025-01-08T04:59:38,851 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithResetTtl 2025-01-08T04:59:38,851 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithResetTtl 2025-01-08T04:59:38,851 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithResetTtl 2025-01-08T04:59:38,851 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithResetTtl 2025-01-08T04:59:38,852 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF 2025-01-08T04:59:38,852 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF 2025-01-08T04:59:38,852 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF 2025-01-08T04:59:38,860 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testExportWithResetTtl' descriptor. 2025-01-08T04:59:38,862 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=62, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testExportWithResetTtl 2025-01-08T04:59:38,862 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testExportWithResetTtl' from region states. 2025-01-08T04:59:38,862 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736312378862"}]},"ts":"9223372036854775807"} 2025-01-08T04:59:38,862 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736312378862"}]},"ts":"9223372036854775807"} 2025-01-08T04:59:38,865 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T04:59:38,865 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => 87d1017402411aba577634d8e1497363, NAME => 'testtb-testExportWithResetTtl,,1736312342809.87d1017402411aba577634d8e1497363.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => ef5fe42efd5a386cbc2993c0042cbfb9, NAME => 'testtb-testExportWithResetTtl,1,1736312342809.ef5fe42efd5a386cbc2993c0042cbfb9.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T04:59:38,865 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testExportWithResetTtl' as deleted. 2025-01-08T04:59:38,865 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736312378865"}]},"ts":"9223372036854775807"} 2025-01-08T04:59:38,868 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testExportWithResetTtl state from META 2025-01-08T04:59:38,874 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportWithResetTtl 2025-01-08T04:59:38,874 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportWithResetTtl 2025-01-08T04:59:38,874 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T04:59:38,874 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T04:59:38,874 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T04:59:38,874 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data null 2025-01-08T04:59:38,875 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportWithResetTtl 2025-01-08T04:59:38,875 INFO [zk-permission-watcher-pool-0 {}] access.AuthManager(135): Skipping permission cache refresh because writable data is empty 2025-01-08T04:59:38,875 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T04:59:38,875 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=62 2025-01-08T04:59:38,937 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(133): Finished pid=62, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportWithResetTtl 2025-01-08T04:59:38,938 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=62, state=SUCCESS; DeleteTableProcedure table=testtb-testExportWithResetTtl in 180 msec 2025-01-08T04:59:38,977 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=62 2025-01-08T04:59:38,977 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testExportWithResetTtl, procId: 62 completed 2025-01-08T04:59:38,986 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "emptySnaptb0-testExportWithResetTtl" 2025-01-08T04:59:38,987 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(380): Deleting snapshot: emptySnaptb0-testExportWithResetTtl 2025-01-08T04:59:38,988 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "snaptb-testExportWithResetTtl" 2025-01-08T04:59:38,990 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb-testExportWithResetTtl 2025-01-08T04:59:38,991 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "snaptb0-testExportWithResetTtl" 2025-01-08T04:59:38,993 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testExportWithResetTtl 2025-01-08T04:59:39,015 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: snapshot.TestMobSecureExportSnapshot#testExportWithResetTtl Thread=779 (was 768) Potentially hanging thread: hconnection-0x168697cd-shared-pool-13 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-6 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x168697cd-shared-pool-12 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-3-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1149249371_22 at /127.0.0.1:55548 [Waiting for operation #5] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (366384994) connection to localhost/127.0.0.1:45939 from jenkins.hfs.0 java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: Thread-2261 java.base@17.0.11/java.io.FileInputStream.readBytes(Native Method) java.base@17.0.11/java.io.FileInputStream.read(FileInputStream.java:276) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:282) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) java.base@17.0.11/sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:281) java.base@17.0.11/sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:324) java.base@17.0.11/sun.nio.cs.StreamDecoder.read(StreamDecoder.java:189) java.base@17.0.11/java.io.InputStreamReader.read(InputStreamReader.java:177) java.base@17.0.11/java.io.BufferedReader.fill(BufferedReader.java:162) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:329) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:396) app//org.apache.hadoop.util.Shell$1.run(Shell.java:1025) Potentially hanging thread: HFileArchiver-5 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x168697cd-shared-pool-11 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-3-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LogDeleter #0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-4 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LogDeleter #0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x168697cd-shared-pool-14 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:37029 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ContainersLauncher #1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:401) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:42391 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: process reaper (pid 10951) java.base@17.0.11/java.lang.ProcessHandleImpl.waitForProcessExit0(Native Method) java.base@17.0.11/java.lang.ProcessHandleImpl$1.run(ProcessHandleImpl.java:150) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (366384994) connection to localhost/127.0.0.1:37029 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1149249371_22 at /127.0.0.1:55062 [Waiting for operation #3] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (366384994) connection to localhost/127.0.0.1:42391 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: hconnection-0x168697cd-shared-pool-16 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x168697cd-shared-pool-15 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-3 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: region-location-4 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1149249371_22 at /127.0.0.1:35338 [Waiting for operation #6] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-6-4 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:62) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:883) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) - Thread LEAK? -, OpenFileDescriptor=803 (was 805), MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=1182 (was 1184), ProcessCount=14 (was 14), AvailableMemoryMB=3957 (was 3757) - AvailableMemoryMB LEAK? - 2025-01-08T04:59:39,015 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=779 is superior to 500 2025-01-08T04:59:39,034 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: snapshot.TestMobSecureExportSnapshot#testExportFileSystemState Thread=779, OpenFileDescriptor=803, MaxFileDescriptor=1048576, SystemLoadAverage=1182, ProcessCount=14, AvailableMemoryMB=3957 2025-01-08T04:59:39,034 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=779 is superior to 500 2025-01-08T04:59:39,036 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.3 create 'testtb-testExportFileSystemState', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T04:59:39,037 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=63, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testExportFileSystemState 2025-01-08T04:59:39,039 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=63, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportFileSystemState execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T04:59:39,039 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "testtb-testExportFileSystemState" procId is: 63 2025-01-08T04:59:39,040 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=63, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testExportFileSystemState execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T04:59:39,040 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=63 2025-01-08T04:59:39,048 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741976_1152 (size=443) 2025-01-08T04:59:39,049 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741976_1152 (size=443) 2025-01-08T04:59:39,049 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741976_1152 (size=443) 2025-01-08T04:59:39,051 INFO [RegionOpenAndInit-testtb-testExportFileSystemState-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 075d25de1c7fc32712f870b2726c6615, NAME => 'testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testtb-testExportFileSystemState', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T04:59:39,052 INFO [RegionOpenAndInit-testtb-testExportFileSystemState-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => 92a7b57994aece1a463017f186215fd9, NAME => 'testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testtb-testExportFileSystemState', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T04:59:39,100 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741977_1153 (size=68) 2025-01-08T04:59:39,101 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741977_1153 (size=68) 2025-01-08T04:59:39,101 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741977_1153 (size=68) 2025-01-08T04:59:39,102 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T04:59:39,102 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-0 {}] regionserver.HRegion(1681): Closing 075d25de1c7fc32712f870b2726c6615, disabling compactions & flushes 2025-01-08T04:59:39,102 INFO [RegionOpenAndInit-testtb-testExportFileSystemState-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615. 2025-01-08T04:59:39,102 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615. 2025-01-08T04:59:39,102 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615. after waiting 0 ms 2025-01-08T04:59:39,103 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615. 2025-01-08T04:59:39,103 INFO [RegionOpenAndInit-testtb-testExportFileSystemState-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615. 2025-01-08T04:59:39,103 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-0 {}] regionserver.HRegion(1635): Region close journal for 075d25de1c7fc32712f870b2726c6615: 2025-01-08T04:59:39,103 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741978_1154 (size=68) 2025-01-08T04:59:39,104 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741978_1154 (size=68) 2025-01-08T04:59:39,104 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741978_1154 (size=68) 2025-01-08T04:59:39,105 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T04:59:39,105 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-1 {}] regionserver.HRegion(1681): Closing 92a7b57994aece1a463017f186215fd9, disabling compactions & flushes 2025-01-08T04:59:39,105 INFO [RegionOpenAndInit-testtb-testExportFileSystemState-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9. 2025-01-08T04:59:39,105 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9. 2025-01-08T04:59:39,105 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9. after waiting 0 ms 2025-01-08T04:59:39,105 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9. 2025-01-08T04:59:39,105 INFO [RegionOpenAndInit-testtb-testExportFileSystemState-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9. 2025-01-08T04:59:39,105 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-1 {}] regionserver.HRegion(1635): Region close journal for 92a7b57994aece1a463017f186215fd9: 2025-01-08T04:59:39,106 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=63, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testExportFileSystemState execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T04:59:39,107 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615.","families":{"info":[{"qualifier":"regioninfo","vlen":67,"tag":[],"timestamp":"1736312379106"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736312379106"}]},"ts":"1736312379106"} 2025-01-08T04:59:39,107 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9.","families":{"info":[{"qualifier":"regioninfo","vlen":67,"tag":[],"timestamp":"1736312379106"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736312379106"}]},"ts":"1736312379106"} 2025-01-08T04:59:39,110 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T04:59:39,111 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=63, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testExportFileSystemState execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T04:59:39,111 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312379111"}]},"ts":"1736312379111"} 2025-01-08T04:59:39,113 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemState, state=ENABLING in hbase:meta 2025-01-08T04:59:39,141 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=63 2025-01-08T04:59:39,221 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(202): Hosts are {f28094b72178=0} racks are {/default-rack=0} 2025-01-08T04:59:39,223 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T04:59:39,223 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T04:59:39,223 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T04:59:39,223 INFO [PEWorker-4 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T04:59:39,223 INFO [PEWorker-4 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T04:59:39,223 INFO [PEWorker-4 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T04:59:39,223 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T04:59:39,223 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=64, ppid=63, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=075d25de1c7fc32712f870b2726c6615, ASSIGN}, {pid=65, ppid=63, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=92a7b57994aece1a463017f186215fd9, ASSIGN}] 2025-01-08T04:59:39,224 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=65, ppid=63, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=92a7b57994aece1a463017f186215fd9, ASSIGN 2025-01-08T04:59:39,225 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=64, ppid=63, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=075d25de1c7fc32712f870b2726c6615, ASSIGN 2025-01-08T04:59:39,225 INFO [PEWorker-1 {}] assignment.TransitRegionStateProcedure(264): Starting pid=64, ppid=63, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=075d25de1c7fc32712f870b2726c6615, ASSIGN; state=OFFLINE, location=f28094b72178,37967,1736312288941; forceNewPlan=false, retain=false 2025-01-08T04:59:39,225 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=65, ppid=63, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=92a7b57994aece1a463017f186215fd9, ASSIGN; state=OFFLINE, location=f28094b72178,37583,1736312288873; forceNewPlan=false, retain=false 2025-01-08T04:59:39,343 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=63 2025-01-08T04:59:39,376 INFO [f28094b72178:42195 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T04:59:39,376 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=64 updating hbase:meta row=075d25de1c7fc32712f870b2726c6615, regionState=OPENING, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T04:59:39,376 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=65 updating hbase:meta row=92a7b57994aece1a463017f186215fd9, regionState=OPENING, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T04:59:39,378 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=66, ppid=65, state=RUNNABLE; OpenRegionProcedure 92a7b57994aece1a463017f186215fd9, server=f28094b72178,37583,1736312288873}] 2025-01-08T04:59:39,380 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=67, ppid=64, state=RUNNABLE; OpenRegionProcedure 075d25de1c7fc32712f870b2726c6615, server=f28094b72178,37967,1736312288941}] 2025-01-08T04:59:39,530 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T04:59:39,532 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T04:59:39,537 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] handler.AssignRegionHandler(135): Open testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615. 2025-01-08T04:59:39,537 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(7285): Opening region: {ENCODED => 075d25de1c7fc32712f870b2726c6615, NAME => 'testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T04:59:39,538 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615. service=AccessControlService 2025-01-08T04:59:39,538 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T04:59:39,539 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportFileSystemState 075d25de1c7fc32712f870b2726c6615 2025-01-08T04:59:39,539 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T04:59:39,539 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(7327): checking encryption for 075d25de1c7fc32712f870b2726c6615 2025-01-08T04:59:39,539 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(7330): checking classloading for 075d25de1c7fc32712f870b2726c6615 2025-01-08T04:59:39,547 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] handler.AssignRegionHandler(135): Open testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9. 2025-01-08T04:59:39,548 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(7285): Opening region: {ENCODED => 92a7b57994aece1a463017f186215fd9, NAME => 'testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T04:59:39,548 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9. service=AccessControlService 2025-01-08T04:59:39,548 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T04:59:39,549 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportFileSystemState 92a7b57994aece1a463017f186215fd9 2025-01-08T04:59:39,549 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T04:59:39,549 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(7327): checking encryption for 92a7b57994aece1a463017f186215fd9 2025-01-08T04:59:39,549 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(7330): checking classloading for 92a7b57994aece1a463017f186215fd9 2025-01-08T04:59:39,552 INFO [StoreOpener-075d25de1c7fc32712f870b2726c6615-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 075d25de1c7fc32712f870b2726c6615 2025-01-08T04:59:39,561 INFO [StoreOpener-075d25de1c7fc32712f870b2726c6615-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 075d25de1c7fc32712f870b2726c6615 columnFamilyName cf 2025-01-08T04:59:39,563 INFO [StoreOpener-92a7b57994aece1a463017f186215fd9-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 92a7b57994aece1a463017f186215fd9 2025-01-08T04:59:39,563 DEBUG [StoreOpener-075d25de1c7fc32712f870b2726c6615-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:59:39,564 INFO [StoreOpener-075d25de1c7fc32712f870b2726c6615-1 {}] regionserver.HStore(327): Store=075d25de1c7fc32712f870b2726c6615/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T04:59:39,565 INFO [StoreOpener-92a7b57994aece1a463017f186215fd9-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 92a7b57994aece1a463017f186215fd9 columnFamilyName cf 2025-01-08T04:59:39,566 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/075d25de1c7fc32712f870b2726c6615 2025-01-08T04:59:39,566 DEBUG [StoreOpener-92a7b57994aece1a463017f186215fd9-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:59:39,566 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/075d25de1c7fc32712f870b2726c6615 2025-01-08T04:59:39,566 INFO [StoreOpener-92a7b57994aece1a463017f186215fd9-1 {}] regionserver.HStore(327): Store=92a7b57994aece1a463017f186215fd9/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T04:59:39,568 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/92a7b57994aece1a463017f186215fd9 2025-01-08T04:59:39,568 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/92a7b57994aece1a463017f186215fd9 2025-01-08T04:59:39,570 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(1085): writing seq id for 075d25de1c7fc32712f870b2726c6615 2025-01-08T04:59:39,572 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(1085): writing seq id for 92a7b57994aece1a463017f186215fd9 2025-01-08T04:59:39,579 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/075d25de1c7fc32712f870b2726c6615/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T04:59:39,581 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(1102): Opened 075d25de1c7fc32712f870b2726c6615; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=62933262, jitterRate=-0.06222131848335266}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T04:59:39,582 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(1001): Region open journal for 075d25de1c7fc32712f870b2726c6615: 2025-01-08T04:59:39,582 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/92a7b57994aece1a463017f186215fd9/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T04:59:39,583 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(1102): Opened 92a7b57994aece1a463017f186215fd9; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=62991933, jitterRate=-0.061347052454948425}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T04:59:39,583 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(1001): Region open journal for 92a7b57994aece1a463017f186215fd9: 2025-01-08T04:59:39,584 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9., pid=66, masterSystemTime=1736312379530 2025-01-08T04:59:39,586 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615., pid=67, masterSystemTime=1736312379532 2025-01-08T04:59:39,591 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9. 2025-01-08T04:59:39,591 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] handler.AssignRegionHandler(164): Opened testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9. 2025-01-08T04:59:39,595 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=65 updating hbase:meta row=92a7b57994aece1a463017f186215fd9, regionState=OPEN, openSeqNum=2, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T04:59:39,596 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615. 2025-01-08T04:59:39,596 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] handler.AssignRegionHandler(164): Opened testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615. 2025-01-08T04:59:39,599 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=64 updating hbase:meta row=075d25de1c7fc32712f870b2726c6615, regionState=OPEN, openSeqNum=2, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T04:59:39,622 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=66, resume processing ppid=65 2025-01-08T04:59:39,625 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=66, ppid=65, state=SUCCESS; OpenRegionProcedure 92a7b57994aece1a463017f186215fd9, server=f28094b72178,37583,1736312288873 in 232 msec 2025-01-08T04:59:39,626 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=65, ppid=63, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=92a7b57994aece1a463017f186215fd9, ASSIGN in 400 msec 2025-01-08T04:59:39,626 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=67, resume processing ppid=64 2025-01-08T04:59:39,626 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=67, ppid=64, state=SUCCESS; OpenRegionProcedure 075d25de1c7fc32712f870b2726c6615, server=f28094b72178,37967,1736312288941 in 239 msec 2025-01-08T04:59:39,637 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=64, resume processing ppid=63 2025-01-08T04:59:39,638 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=64, ppid=63, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=075d25de1c7fc32712f870b2726c6615, ASSIGN in 403 msec 2025-01-08T04:59:39,639 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=63, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testExportFileSystemState execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T04:59:39,639 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312379639"}]},"ts":"1736312379639"} 2025-01-08T04:59:39,643 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemState, state=ENABLED in hbase:meta 2025-01-08T04:59:39,644 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=63 2025-01-08T04:59:39,661 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=63, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportFileSystemState execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T04:59:39,662 DEBUG [PEWorker-4 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testExportFileSystemState jenkins: RWXCA 2025-01-08T04:59:39,665 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37583 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemState], kv [jenkins: RWXCA] 2025-01-08T04:59:39,682 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T04:59:39,682 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T04:59:39,682 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T04:59:39,682 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T04:59:39,698 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemState with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportFileSystemState \x00 \x01 \x02 \x03 \x04 2025-01-08T04:59:39,698 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemState with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportFileSystemState \x00 \x01 \x02 \x03 \x04 2025-01-08T04:59:39,698 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemState with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportFileSystemState \x00 \x01 \x02 \x03 \x04 2025-01-08T04:59:39,699 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemState with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportFileSystemState \x00 \x01 \x02 \x03 \x04 2025-01-08T04:59:39,703 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=63, state=SUCCESS; CreateTableProcedure table=testtb-testExportFileSystemState in 663 msec 2025-01-08T04:59:40,145 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=63 2025-01-08T04:59:40,145 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testExportFileSystemState, procId: 63 completed 2025-01-08T04:59:40,148 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportFileSystemState 2025-01-08T04:59:40,149 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615. 2025-01-08T04:59:40,149 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T04:59:40,166 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } 2025-01-08T04:59:40,166 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736312380166 (current time:1736312380166). 2025-01-08T04:59:40,166 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T04:59:40,166 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot emptySnaptb0-testExportFileSystemState VERSION not specified, setting to 2 2025-01-08T04:59:40,166 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T04:59:40,167 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x1813201b to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@31ac5d46 2025-01-08T04:59:40,191 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@f58845d, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T04:59:40,192 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T04:59:40,193 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:43420, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T04:59:40,195 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x1813201b to 127.0.0.1:59048 2025-01-08T04:59:40,195 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T04:59:40,196 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x04914cbe to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@59f5339d 2025-01-08T04:59:40,245 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@12c43e92, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T04:59:40,247 DEBUG [hconnection-0x23cc4301-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T04:59:40,248 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:43430, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T04:59:40,252 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x04914cbe to 127.0.0.1:59048 2025-01-08T04:59:40,252 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T04:59:40,252 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemState], kv [jenkins: RWXCA] 2025-01-08T04:59:40,253 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T04:59:40,254 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=68, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } 2025-01-08T04:59:40,254 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 }, snapshot procedure id = 68 2025-01-08T04:59:40,255 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T04:59:40,256 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T04:59:40,257 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=68 2025-01-08T04:59:40,291 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T04:59:40,298 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741979_1155 (size=170) 2025-01-08T04:59:40,299 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741979_1155 (size=170) 2025-01-08T04:59:40,299 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741979_1155 (size=170) 2025-01-08T04:59:40,300 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T04:59:40,300 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=69, ppid=68, state=RUNNABLE; SnapshotRegionProcedure 075d25de1c7fc32712f870b2726c6615}, {pid=70, ppid=68, state=RUNNABLE; SnapshotRegionProcedure 92a7b57994aece1a463017f186215fd9}] 2025-01-08T04:59:40,301 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=70, ppid=68, state=RUNNABLE; SnapshotRegionProcedure 92a7b57994aece1a463017f186215fd9 2025-01-08T04:59:40,301 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=69, ppid=68, state=RUNNABLE; SnapshotRegionProcedure 075d25de1c7fc32712f870b2726c6615 2025-01-08T04:59:40,358 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=68 2025-01-08T04:59:40,456 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T04:59:40,456 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T04:59:40,457 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=37967 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=69 2025-01-08T04:59:40,457 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37583 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=70 2025-01-08T04:59:40,457 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=70}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9. 2025-01-08T04:59:40,457 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=69}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615. 2025-01-08T04:59:40,457 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=70}] regionserver.HRegion(2538): Flush status journal for 92a7b57994aece1a463017f186215fd9: 2025-01-08T04:59:40,458 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=70}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9. for emptySnaptb0-testExportFileSystemState completed. 2025-01-08T04:59:40,458 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=69}] regionserver.HRegion(2538): Flush status journal for 075d25de1c7fc32712f870b2726c6615: 2025-01-08T04:59:40,458 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=69}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615. for emptySnaptb0-testExportFileSystemState completed. 2025-01-08T04:59:40,458 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=70}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9.' region-info for snapshot=emptySnaptb0-testExportFileSystemState 2025-01-08T04:59:40,458 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=70}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T04:59:40,458 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=69}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615.' region-info for snapshot=emptySnaptb0-testExportFileSystemState 2025-01-08T04:59:40,458 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=70}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T04:59:40,458 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=69}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T04:59:40,458 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=69}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T04:59:40,551 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741980_1156 (size=71) 2025-01-08T04:59:40,551 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741981_1157 (size=71) 2025-01-08T04:59:40,551 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741981_1157 (size=71) 2025-01-08T04:59:40,551 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741980_1156 (size=71) 2025-01-08T04:59:40,551 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741980_1156 (size=71) 2025-01-08T04:59:40,554 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741981_1157 (size=71) 2025-01-08T04:59:40,559 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=68 2025-01-08T04:59:40,860 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=68 2025-01-08T04:59:40,951 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=70}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9. 2025-01-08T04:59:40,951 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=69}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615. 2025-01-08T04:59:40,952 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=70}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=70 2025-01-08T04:59:40,952 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=69}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=69 2025-01-08T04:59:40,952 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=69 2025-01-08T04:59:40,952 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportFileSystemState on region 075d25de1c7fc32712f870b2726c6615 2025-01-08T04:59:40,952 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=69, ppid=68, state=RUNNABLE; SnapshotRegionProcedure 075d25de1c7fc32712f870b2726c6615 2025-01-08T04:59:40,952 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=70 2025-01-08T04:59:40,953 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportFileSystemState on region 92a7b57994aece1a463017f186215fd9 2025-01-08T04:59:40,953 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=70, ppid=68, state=RUNNABLE; SnapshotRegionProcedure 92a7b57994aece1a463017f186215fd9 2025-01-08T04:59:40,955 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=69, ppid=68, state=SUCCESS; SnapshotRegionProcedure 075d25de1c7fc32712f870b2726c6615 in 653 msec 2025-01-08T04:59:40,956 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=70, resume processing ppid=68 2025-01-08T04:59:40,956 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=70, ppid=68, state=SUCCESS; SnapshotRegionProcedure 92a7b57994aece1a463017f186215fd9 in 653 msec 2025-01-08T04:59:40,956 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T04:59:40,957 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T04:59:40,958 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T04:59:40,959 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T04:59:40,959 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:59:40,959 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(328): No files under family: cf 2025-01-08T04:59:40,984 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741982_1158 (size=63) 2025-01-08T04:59:40,984 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741982_1158 (size=63) 2025-01-08T04:59:40,984 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741982_1158 (size=63) 2025-01-08T04:59:40,989 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T04:59:40,989 DEBUG [PEWorker-2 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for emptySnaptb0-testExportFileSystemState 2025-01-08T04:59:40,990 DEBUG [PEWorker-2 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/emptySnaptb0-testExportFileSystemState 2025-01-08T04:59:41,028 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741983_1159 (size=653) 2025-01-08T04:59:41,028 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741983_1159 (size=653) 2025-01-08T04:59:41,028 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741983_1159 (size=653) 2025-01-08T04:59:41,039 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T04:59:41,051 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T04:59:41,054 DEBUG [PEWorker-2 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/emptySnaptb0-testExportFileSystemState to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/emptySnaptb0-testExportFileSystemState 2025-01-08T04:59:41,061 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T04:59:41,061 DEBUG [PEWorker-2 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 }, snapshot procedure id = 68 2025-01-08T04:59:41,075 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=68, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } in 808 msec 2025-01-08T04:59:41,362 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=68 2025-01-08T04:59:41,362 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportFileSystemState, procId: 68 completed 2025-01-08T04:59:41,365 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37967 {}] regionserver.HRegion(8254): writing data to region testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T04:59:41,366 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37583 {}] regionserver.HRegion(8254): writing data to region testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T04:59:41,370 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportFileSystemState 2025-01-08T04:59:41,370 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615. 2025-01-08T04:59:41,371 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T04:59:41,385 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } 2025-01-08T04:59:41,385 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736312381385 (current time:1736312381385). 2025-01-08T04:59:41,385 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T04:59:41,385 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testExportFileSystemState VERSION not specified, setting to 2 2025-01-08T04:59:41,385 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T04:59:41,386 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x7a45c2da to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@748233e2 2025-01-08T04:59:41,421 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@70bda9aa, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T04:59:41,422 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T04:59:41,424 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:43444, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T04:59:41,425 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x7a45c2da to 127.0.0.1:59048 2025-01-08T04:59:41,425 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T04:59:41,426 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x61b86ea6 to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@4404c50 2025-01-08T04:59:41,483 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@19ad458b, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T04:59:41,484 DEBUG [hconnection-0x5590c696-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T04:59:41,485 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:43452, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T04:59:41,488 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x61b86ea6 to 127.0.0.1:59048 2025-01-08T04:59:41,488 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T04:59:41,488 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemState], kv [jenkins: RWXCA] 2025-01-08T04:59:41,489 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T04:59:41,490 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=71, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } 2025-01-08T04:59:41,490 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 }, snapshot procedure id = 71 2025-01-08T04:59:41,491 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T04:59:41,491 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=71 2025-01-08T04:59:41,492 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T04:59:41,494 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T04:59:41,551 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741984_1160 (size=165) 2025-01-08T04:59:41,551 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741984_1160 (size=165) 2025-01-08T04:59:41,551 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741984_1160 (size=165) 2025-01-08T04:59:41,556 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T04:59:41,556 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=72, ppid=71, state=RUNNABLE; SnapshotRegionProcedure 075d25de1c7fc32712f870b2726c6615}, {pid=73, ppid=71, state=RUNNABLE; SnapshotRegionProcedure 92a7b57994aece1a463017f186215fd9}] 2025-01-08T04:59:41,557 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=73, ppid=71, state=RUNNABLE; SnapshotRegionProcedure 92a7b57994aece1a463017f186215fd9 2025-01-08T04:59:41,557 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=72, ppid=71, state=RUNNABLE; SnapshotRegionProcedure 075d25de1c7fc32712f870b2726c6615 2025-01-08T04:59:41,592 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=71 2025-01-08T04:59:41,701 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportFileSystemState' 2025-01-08T04:59:41,708 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T04:59:41,708 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T04:59:41,708 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0002_000001 (auth:SIMPLE) from 127.0.0.1:37980 2025-01-08T04:59:41,709 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37583 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=73 2025-01-08T04:59:41,709 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37967 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=72 2025-01-08T04:59:41,709 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9. 2025-01-08T04:59:41,709 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615. 2025-01-08T04:59:41,709 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.HRegion(2837): Flushing 92a7b57994aece1a463017f186215fd9 1/1 column families, dataSize=2.93 KB heapSize=6.58 KB 2025-01-08T04:59:41,709 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.HRegion(2837): Flushing 075d25de1c7fc32712f870b2726c6615 1/1 column families, dataSize=333 B heapSize=976 B 2025-01-08T04:59:41,721 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_2/usercache/jenkins/appcache/application_1736312298603_0002/container_1736312298603_0002_01_000001/launch_container.sh] 2025-01-08T04:59:41,721 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_2/usercache/jenkins/appcache/application_1736312298603_0002/container_1736312298603_0002_01_000001/container_tokens] 2025-01-08T04:59:41,721 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_2/usercache/jenkins/appcache/application_1736312298603_0002/container_1736312298603_0002_01_000001/sysfs] 2025-01-08T04:59:41,740 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108fdc919cadc2049248aeb40241fa870e5_92a7b57994aece1a463017f186215fd9 is 71, key is 11f175010e20c627866bbaef0aeedf74/cf:q/1736312381366/Put/seqid=0 2025-01-08T04:59:41,744 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e202501086aa97e6188d841a38656cca645139f50_075d25de1c7fc32712f870b2726c6615 is 71, key is 01ad8d9373300a8e3ff0807c28d26e45/cf:q/1736312381365/Put/seqid=0 2025-01-08T04:59:41,752 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741985_1161 (size=8031) 2025-01-08T04:59:41,752 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741985_1161 (size=8031) 2025-01-08T04:59:41,753 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741985_1161 (size=8031) 2025-01-08T04:59:41,753 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:59:41,759 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108fdc919cadc2049248aeb40241fa870e5_92a7b57994aece1a463017f186215fd9 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/c4ca4238a0b923820dcc509a6f75849b20250108fdc919cadc2049248aeb40241fa870e5_92a7b57994aece1a463017f186215fd9 2025-01-08T04:59:41,760 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/92a7b57994aece1a463017f186215fd9/.tmp/cf/efd7a208df6d47c49e872ce38b3aee17, store: [table=testtb-testExportFileSystemState family=cf region=92a7b57994aece1a463017f186215fd9] 2025-01-08T04:59:41,761 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/92a7b57994aece1a463017f186215fd9/.tmp/cf/efd7a208df6d47c49e872ce38b3aee17 is 209, key is 13108e50830cb9998b9d7fd8437719283/cf:q/1736312381366/Put/seqid=0 2025-01-08T04:59:41,762 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741986_1162 (size=5242) 2025-01-08T04:59:41,762 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741986_1162 (size=5242) 2025-01-08T04:59:41,763 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741986_1162 (size=5242) 2025-01-08T04:59:41,763 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:59:41,769 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e202501086aa97e6188d841a38656cca645139f50_075d25de1c7fc32712f870b2726c6615 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/d41d8cd98f00b204e9800998ecf8427e202501086aa97e6188d841a38656cca645139f50_075d25de1c7fc32712f870b2726c6615 2025-01-08T04:59:41,769 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741987_1163 (size=14587) 2025-01-08T04:59:41,769 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741987_1163 (size=14587) 2025-01-08T04:59:41,770 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741987_1163 (size=14587) 2025-01-08T04:59:41,771 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=2.9 K, hasBloomFilter=true, into tmp file hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/92a7b57994aece1a463017f186215fd9/.tmp/cf/efd7a208df6d47c49e872ce38b3aee17 2025-01-08T04:59:41,771 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/075d25de1c7fc32712f870b2726c6615/.tmp/cf/6685f583d66d48e2a92edd8977f6bb37, store: [table=testtb-testExportFileSystemState family=cf region=075d25de1c7fc32712f870b2726c6615] 2025-01-08T04:59:41,771 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/075d25de1c7fc32712f870b2726c6615/.tmp/cf/6685f583d66d48e2a92edd8977f6bb37 is 209, key is 0e5143cf120e8ceb4e85d0e134ea5cdc7/cf:q/1736312381365/Put/seqid=0 2025-01-08T04:59:41,777 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741988_1164 (size=6328) 2025-01-08T04:59:41,778 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741988_1164 (size=6328) 2025-01-08T04:59:41,779 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741988_1164 (size=6328) 2025-01-08T04:59:41,779 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/92a7b57994aece1a463017f186215fd9/.tmp/cf/efd7a208df6d47c49e872ce38b3aee17 as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/92a7b57994aece1a463017f186215fd9/cf/efd7a208df6d47c49e872ce38b3aee17 2025-01-08T04:59:41,779 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=333, hasBloomFilter=true, into tmp file hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/075d25de1c7fc32712f870b2726c6615/.tmp/cf/6685f583d66d48e2a92edd8977f6bb37 2025-01-08T04:59:41,786 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/92a7b57994aece1a463017f186215fd9/cf/efd7a208df6d47c49e872ce38b3aee17, entries=45, sequenceid=6, filesize=14.2 K 2025-01-08T04:59:41,787 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/075d25de1c7fc32712f870b2726c6615/.tmp/cf/6685f583d66d48e2a92edd8977f6bb37 as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/075d25de1c7fc32712f870b2726c6615/cf/6685f583d66d48e2a92edd8977f6bb37 2025-01-08T04:59:41,787 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.HRegion(3040): Finished flush of dataSize ~2.93 KB/3003, heapSize ~6.56 KB/6720, currentSize=0 B/0 for 92a7b57994aece1a463017f186215fd9 in 78ms, sequenceid=6, compaction requested=false 2025-01-08T04:59:41,787 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.HRegion(2538): Flush status journal for 92a7b57994aece1a463017f186215fd9: 2025-01-08T04:59:41,787 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9. for snaptb0-testExportFileSystemState completed. 2025-01-08T04:59:41,787 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9.' region-info for snapshot=snaptb0-testExportFileSystemState 2025-01-08T04:59:41,788 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T04:59:41,788 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/92a7b57994aece1a463017f186215fd9/cf/efd7a208df6d47c49e872ce38b3aee17] hfiles 2025-01-08T04:59:41,788 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/92a7b57994aece1a463017f186215fd9/cf/efd7a208df6d47c49e872ce38b3aee17 for snapshot=snaptb0-testExportFileSystemState 2025-01-08T04:59:41,792 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/075d25de1c7fc32712f870b2726c6615/cf/6685f583d66d48e2a92edd8977f6bb37, entries=5, sequenceid=6, filesize=6.2 K 2025-01-08T04:59:41,793 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.HRegion(3040): Finished flush of dataSize ~333 B/333, heapSize ~960 B/960, currentSize=0 B/0 for 075d25de1c7fc32712f870b2726c6615 in 84ms, sequenceid=6, compaction requested=false 2025-01-08T04:59:41,793 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=71 2025-01-08T04:59:41,794 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.HRegion(2538): Flush status journal for 075d25de1c7fc32712f870b2726c6615: 2025-01-08T04:59:41,794 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615. for snaptb0-testExportFileSystemState completed. 2025-01-08T04:59:41,794 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615.' region-info for snapshot=snaptb0-testExportFileSystemState 2025-01-08T04:59:41,794 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T04:59:41,794 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/075d25de1c7fc32712f870b2726c6615/cf/6685f583d66d48e2a92edd8977f6bb37] hfiles 2025-01-08T04:59:41,794 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/075d25de1c7fc32712f870b2726c6615/cf/6685f583d66d48e2a92edd8977f6bb37 for snapshot=snaptb0-testExportFileSystemState 2025-01-08T04:59:41,808 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741989_1165 (size=110) 2025-01-08T04:59:41,808 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741989_1165 (size=110) 2025-01-08T04:59:41,809 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741989_1165 (size=110) 2025-01-08T04:59:41,810 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9. 2025-01-08T04:59:41,810 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=73 2025-01-08T04:59:41,810 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=73 2025-01-08T04:59:41,810 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportFileSystemState on region 92a7b57994aece1a463017f186215fd9 2025-01-08T04:59:41,811 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=73, ppid=71, state=RUNNABLE; SnapshotRegionProcedure 92a7b57994aece1a463017f186215fd9 2025-01-08T04:59:41,812 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741990_1166 (size=110) 2025-01-08T04:59:41,813 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741990_1166 (size=110) 2025-01-08T04:59:41,814 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741990_1166 (size=110) 2025-01-08T04:59:41,814 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=73, ppid=71, state=SUCCESS; SnapshotRegionProcedure 92a7b57994aece1a463017f186215fd9 in 256 msec 2025-01-08T04:59:41,814 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615. 2025-01-08T04:59:41,814 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=72 2025-01-08T04:59:41,815 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=72 2025-01-08T04:59:41,815 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportFileSystemState on region 075d25de1c7fc32712f870b2726c6615 2025-01-08T04:59:41,815 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=72, ppid=71, state=RUNNABLE; SnapshotRegionProcedure 075d25de1c7fc32712f870b2726c6615 2025-01-08T04:59:41,817 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=72, resume processing ppid=71 2025-01-08T04:59:41,817 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=72, ppid=71, state=SUCCESS; SnapshotRegionProcedure 075d25de1c7fc32712f870b2726c6615 in 260 msec 2025-01-08T04:59:41,817 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T04:59:41,818 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T04:59:41,819 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T04:59:41,819 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T04:59:41,819 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T04:59:41,821 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/c4ca4238a0b923820dcc509a6f75849b20250108fdc919cadc2049248aeb40241fa870e5_92a7b57994aece1a463017f186215fd9, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/d41d8cd98f00b204e9800998ecf8427e202501086aa97e6188d841a38656cca645139f50_075d25de1c7fc32712f870b2726c6615] hfiles 2025-01-08T04:59:41,821 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/c4ca4238a0b923820dcc509a6f75849b20250108fdc919cadc2049248aeb40241fa870e5_92a7b57994aece1a463017f186215fd9 2025-01-08T04:59:41,821 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/d41d8cd98f00b204e9800998ecf8427e202501086aa97e6188d841a38656cca645139f50_075d25de1c7fc32712f870b2726c6615 2025-01-08T04:59:41,827 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741991_1167 (size=294) 2025-01-08T04:59:41,827 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741991_1167 (size=294) 2025-01-08T04:59:41,827 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741991_1167 (size=294) 2025-01-08T04:59:41,828 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T04:59:41,829 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testExportFileSystemState 2025-01-08T04:59:41,829 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemState 2025-01-08T04:59:41,851 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741992_1168 (size=963) 2025-01-08T04:59:41,851 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741992_1168 (size=963) 2025-01-08T04:59:41,851 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741992_1168 (size=963) 2025-01-08T04:59:41,854 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T04:59:41,861 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T04:59:41,861 DEBUG [PEWorker-1 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemState to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportFileSystemState 2025-01-08T04:59:41,863 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T04:59:41,863 DEBUG [PEWorker-1 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 }, snapshot procedure id = 71 2025-01-08T04:59:41,864 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=71, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } in 374 msec 2025-01-08T04:59:42,095 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=71 2025-01-08T04:59:42,095 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportFileSystemState, procId: 71 completed 2025-01-08T04:59:42,095 INFO [Time-limited test {}] snapshot.TestExportSnapshot(468): HDFS export destination path: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312382095 2025-01-08T04:59:42,095 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=hdfs://localhost:45939, tgtDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312382095, rawTgtDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312382095, srcFsUri=hdfs://localhost:45939, srcDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T04:59:42,131 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:45939, inputRoot=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T04:59:42,131 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1149249371_22, ugi=jenkins (auth:SIMPLE)]], outputRoot=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312382095, skipTmp=false, initialOutputSnapshotDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312382095/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemState 2025-01-08T04:59:42,133 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T04:59:42,141 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportFileSystemState to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312382095/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemState 2025-01-08T04:59:42,176 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741993_1169 (size=165) 2025-01-08T04:59:42,176 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741993_1169 (size=165) 2025-01-08T04:59:42,176 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741993_1169 (size=165) 2025-01-08T04:59:42,179 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741994_1170 (size=963) 2025-01-08T04:59:42,179 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741994_1170 (size=963) 2025-01-08T04:59:42,179 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741994_1170 (size=963) 2025-01-08T04:59:42,577 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.7.0-SNAPSHOT.jar 2025-01-08T04:59:42,578 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T04:59:42,578 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T04:59:42,578 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T04:59:43,398 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T04:59:43,785 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop-11020944075324700641.jar 2025-01-08T04:59:43,785 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T04:59:43,786 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T04:59:43,887 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop-280091847147282115.jar 2025-01-08T04:59:43,888 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T04:59:43,888 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T04:59:43,889 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T04:59:43,889 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T04:59:43,889 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T04:59:43,890 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T04:59:43,890 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T04:59:43,890 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T04:59:43,891 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T04:59:43,891 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T04:59:43,891 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T04:59:43,892 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T04:59:43,892 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T04:59:43,892 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T04:59:43,893 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T04:59:43,893 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T04:59:43,894 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T04:59:43,894 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T04:59:43,895 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T04:59:43,895 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T04:59:43,896 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T04:59:43,896 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T04:59:43,896 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T04:59:43,897 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T04:59:43,897 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T04:59:43,985 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741995_1171 (size=127628) 2025-01-08T04:59:43,985 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741995_1171 (size=127628) 2025-01-08T04:59:43,986 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741995_1171 (size=127628) 2025-01-08T04:59:44,005 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741996_1172 (size=2172137) 2025-01-08T04:59:44,006 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741996_1172 (size=2172137) 2025-01-08T04:59:44,006 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741996_1172 (size=2172137) 2025-01-08T04:59:44,046 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741997_1173 (size=213228) 2025-01-08T04:59:44,047 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741997_1173 (size=213228) 2025-01-08T04:59:44,048 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741997_1173 (size=213228) 2025-01-08T04:59:44,059 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741998_1174 (size=1877034) 2025-01-08T04:59:44,059 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741998_1174 (size=1877034) 2025-01-08T04:59:44,060 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741998_1174 (size=1877034) 2025-01-08T04:59:44,069 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741999_1175 (size=533455) 2025-01-08T04:59:44,069 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741999_1175 (size=533455) 2025-01-08T04:59:44,069 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741999_1175 (size=533455) 2025-01-08T04:59:44,111 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742000_1176 (size=7280644) 2025-01-08T04:59:44,119 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742000_1176 (size=7280644) 2025-01-08T04:59:44,119 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742000_1176 (size=7280644) 2025-01-08T04:59:44,199 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742001_1177 (size=451756) 2025-01-08T04:59:44,200 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742001_1177 (size=451756) 2025-01-08T04:59:44,201 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742001_1177 (size=451756) 2025-01-08T04:59:44,267 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742002_1178 (size=4188619) 2025-01-08T04:59:44,268 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742002_1178 (size=4188619) 2025-01-08T04:59:44,268 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742002_1178 (size=4188619) 2025-01-08T04:59:44,280 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742003_1179 (size=20406) 2025-01-08T04:59:44,281 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742003_1179 (size=20406) 2025-01-08T04:59:44,281 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742003_1179 (size=20406) 2025-01-08T04:59:44,290 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742004_1180 (size=75495) 2025-01-08T04:59:44,290 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742004_1180 (size=75495) 2025-01-08T04:59:44,291 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742004_1180 (size=75495) 2025-01-08T04:59:44,301 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742005_1181 (size=45609) 2025-01-08T04:59:44,301 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742005_1181 (size=45609) 2025-01-08T04:59:44,302 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742005_1181 (size=45609) 2025-01-08T04:59:44,310 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742006_1182 (size=110084) 2025-01-08T04:59:44,310 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742006_1182 (size=110084) 2025-01-08T04:59:44,311 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742006_1182 (size=110084) 2025-01-08T04:59:44,350 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742007_1183 (size=6350712) 2025-01-08T04:59:44,350 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742007_1183 (size=6350712) 2025-01-08T04:59:44,350 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742007_1183 (size=6350712) 2025-01-08T04:59:44,368 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742008_1184 (size=1323991) 2025-01-08T04:59:44,368 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742008_1184 (size=1323991) 2025-01-08T04:59:44,369 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742008_1184 (size=1323991) 2025-01-08T04:59:44,378 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742009_1185 (size=23076) 2025-01-08T04:59:44,378 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742009_1185 (size=23076) 2025-01-08T04:59:44,379 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742009_1185 (size=23076) 2025-01-08T04:59:44,387 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742010_1186 (size=126803) 2025-01-08T04:59:44,387 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742010_1186 (size=126803) 2025-01-08T04:59:44,390 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742010_1186 (size=126803) 2025-01-08T04:59:44,409 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742011_1187 (size=322274) 2025-01-08T04:59:44,410 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742011_1187 (size=322274) 2025-01-08T04:59:44,410 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742011_1187 (size=322274) 2025-01-08T04:59:44,427 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742012_1188 (size=1832290) 2025-01-08T04:59:44,428 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742012_1188 (size=1832290) 2025-01-08T04:59:44,428 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742012_1188 (size=1832290) 2025-01-08T04:59:44,440 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742013_1189 (size=30081) 2025-01-08T04:59:44,440 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742013_1189 (size=30081) 2025-01-08T04:59:44,441 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742013_1189 (size=30081) 2025-01-08T04:59:44,506 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742014_1190 (size=53616) 2025-01-08T04:59:44,506 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742014_1190 (size=53616) 2025-01-08T04:59:44,506 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742014_1190 (size=53616) 2025-01-08T04:59:44,914 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742015_1191 (size=29229) 2025-01-08T04:59:44,914 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742015_1191 (size=29229) 2025-01-08T04:59:44,915 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742015_1191 (size=29229) 2025-01-08T04:59:44,924 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742016_1192 (size=169089) 2025-01-08T04:59:44,924 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742016_1192 (size=169089) 2025-01-08T04:59:44,924 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742016_1192 (size=169089) 2025-01-08T04:59:44,943 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742017_1193 (size=5175431) 2025-01-08T04:59:44,943 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742017_1193 (size=5175431) 2025-01-08T04:59:44,944 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742017_1193 (size=5175431) 2025-01-08T04:59:44,951 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742018_1194 (size=136454) 2025-01-08T04:59:44,952 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742018_1194 (size=136454) 2025-01-08T04:59:44,952 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742018_1194 (size=136454) 2025-01-08T04:59:44,961 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742019_1195 (size=907470) 2025-01-08T04:59:44,961 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742019_1195 (size=907470) 2025-01-08T04:59:44,961 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742019_1195 (size=907470) 2025-01-08T04:59:44,978 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742020_1196 (size=3317408) 2025-01-08T04:59:44,978 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742020_1196 (size=3317408) 2025-01-08T04:59:44,978 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742020_1196 (size=3317408) 2025-01-08T04:59:44,987 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742021_1197 (size=503880) 2025-01-08T04:59:44,988 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742021_1197 (size=503880) 2025-01-08T04:59:44,988 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742021_1197 (size=503880) 2025-01-08T04:59:45,004 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742022_1198 (size=4695811) 2025-01-08T04:59:45,005 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742022_1198 (size=4695811) 2025-01-08T04:59:45,005 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742022_1198 (size=4695811) 2025-01-08T04:59:45,006 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T04:59:45,009 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'snaptb0-testExportFileSystemState' hfile list 2025-01-08T04:59:45,012 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(751): export split=0 size=33.4 K 2025-01-08T04:59:45,025 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742023_1199 (size=726) 2025-01-08T04:59:45,025 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742023_1199 (size=726) 2025-01-08T04:59:45,026 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742023_1199 (size=726) 2025-01-08T04:59:45,038 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742024_1200 (size=15) 2025-01-08T04:59:45,038 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742024_1200 (size=15) 2025-01-08T04:59:45,039 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742024_1200 (size=15) 2025-01-08T04:59:45,057 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742025_1201 (size=304885) 2025-01-08T04:59:45,057 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742025_1201 (size=304885) 2025-01-08T04:59:45,057 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742025_1201 (size=304885) 2025-01-08T04:59:45,075 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T04:59:45,075 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T04:59:45,664 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0003_000001 (auth:SIMPLE) from 127.0.0.1:41354 2025-01-08T04:59:48,345 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemState 2025-01-08T04:59:48,345 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemState Metrics about Tables on a single HBase RegionServer 2025-01-08T04:59:48,346 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportWithResetTtl 2025-01-08T04:59:51,631 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0003_000001 (auth:SIMPLE) from 127.0.0.1:60942 2025-01-08T04:59:51,910 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742026_1202 (size=350559) 2025-01-08T04:59:51,910 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742026_1202 (size=350559) 2025-01-08T04:59:51,910 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742026_1202 (size=350559) 2025-01-08T04:59:53,850 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T04:59:54,027 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0003_000001 (auth:SIMPLE) from 127.0.0.1:36444 2025-01-08T05:00:00,139 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742027_1203 (size=14587) 2025-01-08T05:00:00,139 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742027_1203 (size=14587) 2025-01-08T05:00:00,139 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742027_1203 (size=14587) 2025-01-08T05:00:00,193 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742028_1204 (size=8031) 2025-01-08T05:00:00,193 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742028_1204 (size=8031) 2025-01-08T05:00:00,193 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742028_1204 (size=8031) 2025-01-08T05:00:00,248 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742029_1205 (size=6328) 2025-01-08T05:00:00,248 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742029_1205 (size=6328) 2025-01-08T05:00:00,248 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742029_1205 (size=6328) 2025-01-08T05:00:00,295 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742030_1206 (size=5242) 2025-01-08T05:00:00,295 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742030_1206 (size=5242) 2025-01-08T05:00:00,295 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742030_1206 (size=5242) 2025-01-08T05:00:00,383 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742031_1207 (size=17462) 2025-01-08T05:00:00,383 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742031_1207 (size=17462) 2025-01-08T05:00:00,384 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742031_1207 (size=17462) 2025-01-08T05:00:00,403 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742032_1208 (size=465) 2025-01-08T05:00:00,403 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742032_1208 (size=465) 2025-01-08T05:00:00,403 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742032_1208 (size=465) 2025-01-08T05:00:00,456 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742033_1209 (size=17462) 2025-01-08T05:00:00,457 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742033_1209 (size=17462) 2025-01-08T05:00:00,457 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742033_1209 (size=17462) 2025-01-08T05:00:00,470 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_1/usercache/jenkins/appcache/application_1736312298603_0003/container_1736312298603_0003_01_000002/launch_container.sh] 2025-01-08T05:00:00,471 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_1/usercache/jenkins/appcache/application_1736312298603_0003/container_1736312298603_0003_01_000002/container_tokens] 2025-01-08T05:00:00,471 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_1/usercache/jenkins/appcache/application_1736312298603_0003/container_1736312298603_0003_01_000002/sysfs] 2025-01-08T05:00:00,480 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742034_1210 (size=350559) 2025-01-08T05:00:00,480 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742034_1210 (size=350559) 2025-01-08T05:00:00,480 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742034_1210 (size=350559) 2025-01-08T05:00:00,501 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0003_000001 (auth:SIMPLE) from 127.0.0.1:44744 2025-01-08T05:00:02,265 INFO [Time-limited test {}] snapshot.ExportSnapshot(1207): Finalize the Snapshot Export 2025-01-08T05:00:02,268 INFO [Time-limited test {}] snapshot.ExportSnapshot(1218): Verify the exported snapshot's expiration status and integrity. 2025-01-08T05:00:02,280 INFO [Time-limited test {}] snapshot.ExportSnapshot(1224): Export Completed: snaptb0-testExportFileSystemState 2025-01-08T05:00:02,280 INFO [Time-limited test {}] snapshot.TestExportSnapshot(362): Exported snapshot 2025-01-08T05:00:02,281 INFO [Time-limited test {}] snapshot.TestExportSnapshot(373): Verified filesystem state 2025-01-08T05:00:02,281 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1149249371_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportFileSystemState at hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportFileSystemState 2025-01-08T05:00:02,282 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportFileSystemState/.snapshotinfo 2025-01-08T05:00:02,282 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportFileSystemState/data.manifest 2025-01-08T05:00:02,282 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1149249371_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312382095/.hbase-snapshot/snaptb0-testExportFileSystemState at hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312382095/.hbase-snapshot/snaptb0-testExportFileSystemState 2025-01-08T05:00:02,282 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312382095/.hbase-snapshot/snaptb0-testExportFileSystemState/.snapshotinfo 2025-01-08T05:00:02,282 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312382095/.hbase-snapshot/snaptb0-testExportFileSystemState/data.manifest 2025-01-08T05:00:02,295 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testExportFileSystemState 2025-01-08T05:00:02,296 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.3 disable testtb-testExportFileSystemState 2025-01-08T05:00:02,303 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=74, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testExportFileSystemState 2025-01-08T05:00:02,308 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=74 2025-01-08T05:00:02,309 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312402309"}]},"ts":"1736312402309"} 2025-01-08T05:00:02,311 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemState, state=DISABLING in hbase:meta 2025-01-08T05:00:02,322 INFO [PEWorker-3 {}] procedure.DisableTableProcedure(284): Set testtb-testExportFileSystemState to state=DISABLING 2025-01-08T05:00:02,323 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=75, ppid=74, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testExportFileSystemState}] 2025-01-08T05:00:02,325 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=76, ppid=75, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=075d25de1c7fc32712f870b2726c6615, UNASSIGN}, {pid=77, ppid=75, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=92a7b57994aece1a463017f186215fd9, UNASSIGN}] 2025-01-08T05:00:02,326 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=77, ppid=75, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=92a7b57994aece1a463017f186215fd9, UNASSIGN 2025-01-08T05:00:02,326 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=76, ppid=75, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=075d25de1c7fc32712f870b2726c6615, UNASSIGN 2025-01-08T05:00:02,327 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=76 updating hbase:meta row=075d25de1c7fc32712f870b2726c6615, regionState=CLOSING, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T05:00:02,327 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=77 updating hbase:meta row=92a7b57994aece1a463017f186215fd9, regionState=CLOSING, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T05:00:02,330 DEBUG [PEWorker-4 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T05:00:02,330 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=78, ppid=77, state=RUNNABLE; CloseRegionProcedure 92a7b57994aece1a463017f186215fd9, server=f28094b72178,37583,1736312288873}] 2025-01-08T05:00:02,331 DEBUG [PEWorker-2 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T05:00:02,332 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=79, ppid=76, state=RUNNABLE; CloseRegionProcedure 075d25de1c7fc32712f870b2726c6615, server=f28094b72178,37967,1736312288941}] 2025-01-08T05:00:02,412 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=74 2025-01-08T05:00:02,486 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:00:02,486 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T05:00:02,487 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] handler.UnassignRegionHandler(124): Close 92a7b57994aece1a463017f186215fd9 2025-01-08T05:00:02,488 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T05:00:02,488 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] regionserver.HRegion(1681): Closing 92a7b57994aece1a463017f186215fd9, disabling compactions & flushes 2025-01-08T05:00:02,488 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9. 2025-01-08T05:00:02,488 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9. 2025-01-08T05:00:02,488 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9. after waiting 0 ms 2025-01-08T05:00:02,488 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9. 2025-01-08T05:00:02,490 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] handler.UnassignRegionHandler(124): Close 075d25de1c7fc32712f870b2726c6615 2025-01-08T05:00:02,490 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T05:00:02,491 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] regionserver.HRegion(1681): Closing 075d25de1c7fc32712f870b2726c6615, disabling compactions & flushes 2025-01-08T05:00:02,491 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615. 2025-01-08T05:00:02,491 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615. 2025-01-08T05:00:02,491 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615. after waiting 0 ms 2025-01-08T05:00:02,491 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615. 2025-01-08T05:00:02,499 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/075d25de1c7fc32712f870b2726c6615/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T05:00:02,499 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/92a7b57994aece1a463017f186215fd9/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T05:00:02,500 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:00:02,500 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:00:02,500 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615. 2025-01-08T05:00:02,500 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] regionserver.HRegion(1635): Region close journal for 075d25de1c7fc32712f870b2726c6615: 2025-01-08T05:00:02,500 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9. 2025-01-08T05:00:02,500 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] regionserver.HRegion(1635): Region close journal for 92a7b57994aece1a463017f186215fd9: 2025-01-08T05:00:02,509 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] handler.UnassignRegionHandler(170): Closed 075d25de1c7fc32712f870b2726c6615 2025-01-08T05:00:02,509 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=76 updating hbase:meta row=075d25de1c7fc32712f870b2726c6615, regionState=CLOSED 2025-01-08T05:00:02,511 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] handler.UnassignRegionHandler(170): Closed 92a7b57994aece1a463017f186215fd9 2025-01-08T05:00:02,512 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=77 updating hbase:meta row=92a7b57994aece1a463017f186215fd9, regionState=CLOSED 2025-01-08T05:00:02,522 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=79, resume processing ppid=76 2025-01-08T05:00:02,523 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=79, ppid=76, state=SUCCESS; CloseRegionProcedure 075d25de1c7fc32712f870b2726c6615, server=f28094b72178,37967,1736312288941 in 189 msec 2025-01-08T05:00:02,523 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=76, ppid=75, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=075d25de1c7fc32712f870b2726c6615, UNASSIGN in 197 msec 2025-01-08T05:00:02,524 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=78, resume processing ppid=77 2025-01-08T05:00:02,525 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=78, ppid=77, state=SUCCESS; CloseRegionProcedure 92a7b57994aece1a463017f186215fd9, server=f28094b72178,37583,1736312288873 in 193 msec 2025-01-08T05:00:02,527 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=77, resume processing ppid=75 2025-01-08T05:00:02,527 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=77, ppid=75, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=92a7b57994aece1a463017f186215fd9, UNASSIGN in 200 msec 2025-01-08T05:00:02,531 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=75, resume processing ppid=74 2025-01-08T05:00:02,531 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=75, ppid=74, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testExportFileSystemState in 207 msec 2025-01-08T05:00:02,541 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312402541"}]},"ts":"1736312402541"} 2025-01-08T05:00:02,543 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemState, state=DISABLED in hbase:meta 2025-01-08T05:00:02,551 INFO [PEWorker-4 {}] procedure.DisableTableProcedure(296): Set testtb-testExportFileSystemState to state=DISABLED 2025-01-08T05:00:02,554 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=74, state=SUCCESS; DisableTableProcedure table=testtb-testExportFileSystemState in 256 msec 2025-01-08T05:00:02,613 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=74 2025-01-08T05:00:02,614 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testExportFileSystemState, procId: 74 completed 2025-01-08T05:00:02,614 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.3 delete testtb-testExportFileSystemState 2025-01-08T05:00:02,615 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=80, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testExportFileSystemState 2025-01-08T05:00:02,616 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=80, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemState 2025-01-08T05:00:02,616 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testExportFileSystemState 2025-01-08T05:00:02,617 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=80, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemState 2025-01-08T05:00:02,619 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37583 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testExportFileSystemState 2025-01-08T05:00:02,621 DEBUG [HFileArchiver-7 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/075d25de1c7fc32712f870b2726c6615 2025-01-08T05:00:02,621 DEBUG [HFileArchiver-8 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/92a7b57994aece1a463017f186215fd9 2025-01-08T05:00:02,623 DEBUG [HFileArchiver-8 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/92a7b57994aece1a463017f186215fd9/cf, FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/92a7b57994aece1a463017f186215fd9/recovered.edits] 2025-01-08T05:00:02,623 DEBUG [HFileArchiver-7 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/075d25de1c7fc32712f870b2726c6615/cf, FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/075d25de1c7fc32712f870b2726c6615/recovered.edits] 2025-01-08T05:00:02,644 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemState 2025-01-08T05:00:02,644 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemState 2025-01-08T05:00:02,644 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemState 2025-01-08T05:00:02,644 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemState 2025-01-08T05:00:02,646 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemState with data PBUF 2025-01-08T05:00:02,646 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemState with data PBUF 2025-01-08T05:00:02,646 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemState with data PBUF 2025-01-08T05:00:02,650 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemState with data PBUF 2025-01-08T05:00:02,650 DEBUG [HFileArchiver-8 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/92a7b57994aece1a463017f186215fd9/cf/efd7a208df6d47c49e872ce38b3aee17 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportFileSystemState/92a7b57994aece1a463017f186215fd9/cf/efd7a208df6d47c49e872ce38b3aee17 2025-01-08T05:00:02,650 DEBUG [HFileArchiver-7 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/075d25de1c7fc32712f870b2726c6615/cf/6685f583d66d48e2a92edd8977f6bb37 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportFileSystemState/075d25de1c7fc32712f870b2726c6615/cf/6685f583d66d48e2a92edd8977f6bb37 2025-01-08T05:00:02,655 DEBUG [HFileArchiver-7 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/075d25de1c7fc32712f870b2726c6615/recovered.edits/9.seqid to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportFileSystemState/075d25de1c7fc32712f870b2726c6615/recovered.edits/9.seqid 2025-01-08T05:00:02,656 DEBUG [HFileArchiver-7 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/075d25de1c7fc32712f870b2726c6615 2025-01-08T05:00:02,662 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemState 2025-01-08T05:00:02,663 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemState 2025-01-08T05:00:02,663 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:00:02,663 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:00:02,666 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemState 2025-01-08T05:00:02,666 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:00:02,666 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemState 2025-01-08T05:00:02,666 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:00:02,668 DEBUG [HFileArchiver-8 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/92a7b57994aece1a463017f186215fd9/recovered.edits/9.seqid to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportFileSystemState/92a7b57994aece1a463017f186215fd9/recovered.edits/9.seqid 2025-01-08T05:00:02,669 DEBUG [HFileArchiver-8 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemState/92a7b57994aece1a463017f186215fd9 2025-01-08T05:00:02,669 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(313): Archived testtb-testExportFileSystemState regions 2025-01-08T05:00:02,670 DEBUG [PEWorker-3 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d 2025-01-08T05:00:02,671 DEBUG [PEWorker-3 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf] 2025-01-08T05:00:02,671 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=80 2025-01-08T05:00:02,676 DEBUG [PEWorker-3 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/c4ca4238a0b923820dcc509a6f75849b20250108fdc919cadc2049248aeb40241fa870e5_92a7b57994aece1a463017f186215fd9 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/c4ca4238a0b923820dcc509a6f75849b20250108fdc919cadc2049248aeb40241fa870e5_92a7b57994aece1a463017f186215fd9 2025-01-08T05:00:02,678 DEBUG [PEWorker-3 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/d41d8cd98f00b204e9800998ecf8427e202501086aa97e6188d841a38656cca645139f50_075d25de1c7fc32712f870b2726c6615 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/d41d8cd98f00b204e9800998ecf8427e202501086aa97e6188d841a38656cca645139f50_075d25de1c7fc32712f870b2726c6615 2025-01-08T05:00:02,679 DEBUG [PEWorker-3 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d 2025-01-08T05:00:02,685 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=80, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemState 2025-01-08T05:00:02,707 WARN [PEWorker-3 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testtb-testExportFileSystemState from hbase:meta 2025-01-08T05:00:02,714 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testExportFileSystemState' descriptor. 2025-01-08T05:00:02,716 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=80, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemState 2025-01-08T05:00:02,716 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testExportFileSystemState' from region states. 2025-01-08T05:00:02,717 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736312402716"}]},"ts":"9223372036854775807"} 2025-01-08T05:00:02,717 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736312402716"}]},"ts":"9223372036854775807"} 2025-01-08T05:00:02,737 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T05:00:02,737 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => 075d25de1c7fc32712f870b2726c6615, NAME => 'testtb-testExportFileSystemState,,1736312379036.075d25de1c7fc32712f870b2726c6615.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => 92a7b57994aece1a463017f186215fd9, NAME => 'testtb-testExportFileSystemState,1,1736312379036.92a7b57994aece1a463017f186215fd9.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T05:00:02,737 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testExportFileSystemState' as deleted. 2025-01-08T05:00:02,737 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736312402737"}]},"ts":"9223372036854775807"} 2025-01-08T05:00:02,746 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testExportFileSystemState state from META 2025-01-08T05:00:02,767 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(133): Finished pid=80, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemState 2025-01-08T05:00:02,772 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=80 2025-01-08T05:00:02,781 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=80, state=SUCCESS; DeleteTableProcedure table=testtb-testExportFileSystemState in 153 msec 2025-01-08T05:00:02,974 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=80 2025-01-08T05:00:02,974 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testExportFileSystemState, procId: 80 completed 2025-01-08T05:00:02,983 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "emptySnaptb0-testExportFileSystemState" 2025-01-08T05:00:02,985 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotManager(380): Deleting snapshot: emptySnaptb0-testExportFileSystemState 2025-01-08T05:00:02,986 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "snaptb0-testExportFileSystemState" 2025-01-08T05:00:02,988 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testExportFileSystemState 2025-01-08T05:00:03,019 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: snapshot.TestMobSecureExportSnapshot#testExportFileSystemState Thread=786 (was 779) Potentially hanging thread: IPC Client (366384994) connection to localhost/127.0.0.1:37123 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: hconnection-0x168697cd-shared-pool-18 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x168697cd-shared-pool-19 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LogDeleter #1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1177) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1297705046_1 at /127.0.0.1:36988 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Thread-2938 java.base@17.0.11/java.io.FileInputStream.readBytes(Native Method) java.base@17.0.11/java.io.FileInputStream.read(FileInputStream.java:276) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:282) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) java.base@17.0.11/sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:281) java.base@17.0.11/sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:324) java.base@17.0.11/sun.nio.cs.StreamDecoder.read(StreamDecoder.java:189) java.base@17.0.11/java.io.InputStreamReader.read(InputStreamReader.java:177) java.base@17.0.11/java.io.BufferedReader.fill(BufferedReader.java:162) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:329) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:396) app//org.apache.hadoop.util.Shell$1.run(Shell.java:1025) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1149249371_22 at /127.0.0.1:48308 [Waiting for operation #6] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1297705046_1 at /127.0.0.1:48300 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-7 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: process reaper (pid 12740) java.base@17.0.11/java.lang.ProcessHandleImpl.waitForProcessExit0(Native Method) java.base@17.0.11/java.lang.ProcessHandleImpl$1.run(ProcessHandleImpl.java:150) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x168697cd-shared-pool-17 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-8 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:45809 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #4 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1149249371_22 at /127.0.0.1:36998 [Waiting for operation #5] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x168697cd-shared-pool-20 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #3 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:37123 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) - Thread LEAK? -, OpenFileDescriptor=803 (was 803), MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=1141 (was 1182), ProcessCount=17 (was 14) - ProcessCount LEAK? -, AvailableMemoryMB=3994 (was 3957) - AvailableMemoryMB LEAK? - 2025-01-08T05:00:03,019 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=786 is superior to 500 2025-01-08T05:00:03,038 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: snapshot.TestMobSecureExportSnapshot#testConsecutiveExports Thread=786, OpenFileDescriptor=803, MaxFileDescriptor=1048576, SystemLoadAverage=1141, ProcessCount=17, AvailableMemoryMB=3993 2025-01-08T05:00:03,039 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=786 is superior to 500 2025-01-08T05:00:03,041 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.3 create 'testtb-testConsecutiveExports', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T05:00:03,042 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=81, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testConsecutiveExports 2025-01-08T05:00:03,045 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=81, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testConsecutiveExports execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T05:00:03,045 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "testtb-testConsecutiveExports" procId is: 81 2025-01-08T05:00:03,046 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=81 2025-01-08T05:00:03,047 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=81, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testConsecutiveExports execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T05:00:03,076 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742035_1211 (size=440) 2025-01-08T05:00:03,076 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742035_1211 (size=440) 2025-01-08T05:00:03,077 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742035_1211 (size=440) 2025-01-08T05:00:03,079 INFO [RegionOpenAndInit-testtb-testConsecutiveExports-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => b92a3e58f76a25816db26a2020c9df1b, NAME => 'testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testtb-testConsecutiveExports', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:00:03,080 INFO [RegionOpenAndInit-testtb-testConsecutiveExports-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => c432634a817b2f286fb158c1aad95aab, NAME => 'testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testtb-testConsecutiveExports', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:00:03,120 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742036_1212 (size=65) 2025-01-08T05:00:03,120 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742037_1213 (size=65) 2025-01-08T05:00:03,121 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742036_1212 (size=65) 2025-01-08T05:00:03,122 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742036_1212 (size=65) 2025-01-08T05:00:03,122 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742037_1213 (size=65) 2025-01-08T05:00:03,122 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:00:03,122 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742037_1213 (size=65) 2025-01-08T05:00:03,122 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-1 {}] regionserver.HRegion(1681): Closing c432634a817b2f286fb158c1aad95aab, disabling compactions & flushes 2025-01-08T05:00:03,123 INFO [RegionOpenAndInit-testtb-testConsecutiveExports-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab. 2025-01-08T05:00:03,123 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab. 2025-01-08T05:00:03,123 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab. after waiting 0 ms 2025-01-08T05:00:03,123 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab. 2025-01-08T05:00:03,123 INFO [RegionOpenAndInit-testtb-testConsecutiveExports-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab. 2025-01-08T05:00:03,123 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-1 {}] regionserver.HRegion(1635): Region close journal for c432634a817b2f286fb158c1aad95aab: 2025-01-08T05:00:03,123 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:00:03,123 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-0 {}] regionserver.HRegion(1681): Closing b92a3e58f76a25816db26a2020c9df1b, disabling compactions & flushes 2025-01-08T05:00:03,123 INFO [RegionOpenAndInit-testtb-testConsecutiveExports-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b. 2025-01-08T05:00:03,123 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b. 2025-01-08T05:00:03,123 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b. after waiting 0 ms 2025-01-08T05:00:03,123 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b. 2025-01-08T05:00:03,123 INFO [RegionOpenAndInit-testtb-testConsecutiveExports-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b. 2025-01-08T05:00:03,123 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-0 {}] regionserver.HRegion(1635): Region close journal for b92a3e58f76a25816db26a2020c9df1b: 2025-01-08T05:00:03,125 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=81, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testConsecutiveExports execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T05:00:03,125 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab.","families":{"info":[{"qualifier":"regioninfo","vlen":64,"tag":[],"timestamp":"1736312403125"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736312403125"}]},"ts":"1736312403125"} 2025-01-08T05:00:03,125 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b.","families":{"info":[{"qualifier":"regioninfo","vlen":64,"tag":[],"timestamp":"1736312403125"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736312403125"}]},"ts":"1736312403125"} 2025-01-08T05:00:03,128 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T05:00:03,129 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=81, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testConsecutiveExports execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T05:00:03,130 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testConsecutiveExports","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312403130"}]},"ts":"1736312403130"} 2025-01-08T05:00:03,132 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testConsecutiveExports, state=ENABLING in hbase:meta 2025-01-08T05:00:03,152 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=81 2025-01-08T05:00:03,167 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(202): Hosts are {f28094b72178=0} racks are {/default-rack=0} 2025-01-08T05:00:03,169 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T05:00:03,169 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T05:00:03,169 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T05:00:03,169 INFO [PEWorker-1 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T05:00:03,169 INFO [PEWorker-1 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T05:00:03,169 INFO [PEWorker-1 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T05:00:03,169 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T05:00:03,170 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=82, ppid=81, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=b92a3e58f76a25816db26a2020c9df1b, ASSIGN}, {pid=83, ppid=81, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=c432634a817b2f286fb158c1aad95aab, ASSIGN}] 2025-01-08T05:00:03,177 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=83, ppid=81, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=c432634a817b2f286fb158c1aad95aab, ASSIGN 2025-01-08T05:00:03,177 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=82, ppid=81, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=b92a3e58f76a25816db26a2020c9df1b, ASSIGN 2025-01-08T05:00:03,181 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=83, ppid=81, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=c432634a817b2f286fb158c1aad95aab, ASSIGN; state=OFFLINE, location=f28094b72178,39801,1736312288779; forceNewPlan=false, retain=false 2025-01-08T05:00:03,181 INFO [PEWorker-5 {}] assignment.TransitRegionStateProcedure(264): Starting pid=82, ppid=81, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=b92a3e58f76a25816db26a2020c9df1b, ASSIGN; state=OFFLINE, location=f28094b72178,37967,1736312288941; forceNewPlan=false, retain=false 2025-01-08T05:00:03,332 INFO [f28094b72178:42195 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T05:00:03,332 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=82 updating hbase:meta row=b92a3e58f76a25816db26a2020c9df1b, regionState=OPENING, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T05:00:03,332 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=83 updating hbase:meta row=c432634a817b2f286fb158c1aad95aab, regionState=OPENING, regionLocation=f28094b72178,39801,1736312288779 2025-01-08T05:00:03,335 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=84, ppid=82, state=RUNNABLE; OpenRegionProcedure b92a3e58f76a25816db26a2020c9df1b, server=f28094b72178,37967,1736312288941}] 2025-01-08T05:00:03,336 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=85, ppid=83, state=RUNNABLE; OpenRegionProcedure c432634a817b2f286fb158c1aad95aab, server=f28094b72178,39801,1736312288779}] 2025-01-08T05:00:03,354 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=81 2025-01-08T05:00:03,488 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:00:03,490 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,39801,1736312288779 2025-01-08T05:00:03,494 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] handler.AssignRegionHandler(135): Open testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b. 2025-01-08T05:00:03,494 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(7285): Opening region: {ENCODED => b92a3e58f76a25816db26a2020c9df1b, NAME => 'testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T05:00:03,495 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b. service=AccessControlService 2025-01-08T05:00:03,495 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T05:00:03,496 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testConsecutiveExports b92a3e58f76a25816db26a2020c9df1b 2025-01-08T05:00:03,496 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(894): Instantiated testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:00:03,496 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(7327): checking encryption for b92a3e58f76a25816db26a2020c9df1b 2025-01-08T05:00:03,496 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(7330): checking classloading for b92a3e58f76a25816db26a2020c9df1b 2025-01-08T05:00:03,500 INFO [StoreOpener-b92a3e58f76a25816db26a2020c9df1b-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region b92a3e58f76a25816db26a2020c9df1b 2025-01-08T05:00:03,502 INFO [StoreOpener-b92a3e58f76a25816db26a2020c9df1b-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region b92a3e58f76a25816db26a2020c9df1b columnFamilyName cf 2025-01-08T05:00:03,504 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] handler.AssignRegionHandler(135): Open testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab. 2025-01-08T05:00:03,504 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(7285): Opening region: {ENCODED => c432634a817b2f286fb158c1aad95aab, NAME => 'testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T05:00:03,504 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab. service=AccessControlService 2025-01-08T05:00:03,504 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T05:00:03,505 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testConsecutiveExports c432634a817b2f286fb158c1aad95aab 2025-01-08T05:00:03,505 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(894): Instantiated testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:00:03,505 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(7327): checking encryption for c432634a817b2f286fb158c1aad95aab 2025-01-08T05:00:03,505 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(7330): checking classloading for c432634a817b2f286fb158c1aad95aab 2025-01-08T05:00:03,510 DEBUG [StoreOpener-b92a3e58f76a25816db26a2020c9df1b-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:00:03,514 INFO [StoreOpener-b92a3e58f76a25816db26a2020c9df1b-1 {}] regionserver.HStore(327): Store=b92a3e58f76a25816db26a2020c9df1b/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T05:00:03,518 INFO [StoreOpener-c432634a817b2f286fb158c1aad95aab-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region c432634a817b2f286fb158c1aad95aab 2025-01-08T05:00:03,519 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/b92a3e58f76a25816db26a2020c9df1b 2025-01-08T05:00:03,520 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/b92a3e58f76a25816db26a2020c9df1b 2025-01-08T05:00:03,526 INFO [StoreOpener-c432634a817b2f286fb158c1aad95aab-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region c432634a817b2f286fb158c1aad95aab columnFamilyName cf 2025-01-08T05:00:03,526 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(1085): writing seq id for b92a3e58f76a25816db26a2020c9df1b 2025-01-08T05:00:03,527 DEBUG [StoreOpener-c432634a817b2f286fb158c1aad95aab-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:00:03,528 INFO [StoreOpener-c432634a817b2f286fb158c1aad95aab-1 {}] regionserver.HStore(327): Store=c432634a817b2f286fb158c1aad95aab/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T05:00:03,530 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/b92a3e58f76a25816db26a2020c9df1b/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T05:00:03,531 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(1102): Opened b92a3e58f76a25816db26a2020c9df1b; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=73381192, jitterRate=0.09346497058868408}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T05:00:03,532 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(1001): Region open journal for b92a3e58f76a25816db26a2020c9df1b: 2025-01-08T05:00:03,533 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/c432634a817b2f286fb158c1aad95aab 2025-01-08T05:00:03,533 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b., pid=84, masterSystemTime=1736312403488 2025-01-08T05:00:03,533 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/c432634a817b2f286fb158c1aad95aab 2025-01-08T05:00:03,535 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b. 2025-01-08T05:00:03,535 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] handler.AssignRegionHandler(164): Opened testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b. 2025-01-08T05:00:03,538 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(1085): writing seq id for c432634a817b2f286fb158c1aad95aab 2025-01-08T05:00:03,539 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=82 updating hbase:meta row=b92a3e58f76a25816db26a2020c9df1b, regionState=OPEN, openSeqNum=2, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T05:00:03,545 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=84, resume processing ppid=82 2025-01-08T05:00:03,545 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=84, ppid=82, state=SUCCESS; OpenRegionProcedure b92a3e58f76a25816db26a2020c9df1b, server=f28094b72178,37967,1736312288941 in 207 msec 2025-01-08T05:00:03,550 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/c432634a817b2f286fb158c1aad95aab/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T05:00:03,551 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(1102): Opened c432634a817b2f286fb158c1aad95aab; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=74380942, jitterRate=0.10836240649223328}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T05:00:03,551 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(1001): Region open journal for c432634a817b2f286fb158c1aad95aab: 2025-01-08T05:00:03,553 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab., pid=85, masterSystemTime=1736312403490 2025-01-08T05:00:03,553 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=82, ppid=81, state=SUCCESS; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=b92a3e58f76a25816db26a2020c9df1b, ASSIGN in 376 msec 2025-01-08T05:00:03,555 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab. 2025-01-08T05:00:03,555 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] handler.AssignRegionHandler(164): Opened testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab. 2025-01-08T05:00:03,556 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=83 updating hbase:meta row=c432634a817b2f286fb158c1aad95aab, regionState=OPEN, openSeqNum=2, regionLocation=f28094b72178,39801,1736312288779 2025-01-08T05:00:03,563 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=85, resume processing ppid=83 2025-01-08T05:00:03,563 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=85, ppid=83, state=SUCCESS; OpenRegionProcedure c432634a817b2f286fb158c1aad95aab, server=f28094b72178,39801,1736312288779 in 222 msec 2025-01-08T05:00:03,565 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=83, resume processing ppid=81 2025-01-08T05:00:03,566 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=83, ppid=81, state=SUCCESS; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=c432634a817b2f286fb158c1aad95aab, ASSIGN in 394 msec 2025-01-08T05:00:03,567 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=81, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testConsecutiveExports execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T05:00:03,568 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testConsecutiveExports","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312403567"}]},"ts":"1736312403567"} 2025-01-08T05:00:03,574 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testConsecutiveExports, state=ENABLED in hbase:meta 2025-01-08T05:00:03,610 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=81, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testConsecutiveExports execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T05:00:03,611 DEBUG [PEWorker-5 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testConsecutiveExports jenkins: RWXCA 2025-01-08T05:00:03,619 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37583 {}] access.PermissionStorage(611): Read acl: entry[testtb-testConsecutiveExports], kv [jenkins: RWXCA] 2025-01-08T05:00:03,628 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:00:03,628 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:00:03,628 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:00:03,629 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:00:03,656 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=81 2025-01-08T05:00:03,660 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testConsecutiveExports with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testConsecutiveExports \x00 \x01 \x02 \x03 \x04 2025-01-08T05:00:03,660 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testConsecutiveExports with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testConsecutiveExports \x00 \x01 \x02 \x03 \x04 2025-01-08T05:00:03,660 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testConsecutiveExports with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testConsecutiveExports \x00 \x01 \x02 \x03 \x04 2025-01-08T05:00:03,660 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testConsecutiveExports with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testConsecutiveExports \x00 \x01 \x02 \x03 \x04 2025-01-08T05:00:03,671 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=81, state=SUCCESS; CreateTableProcedure table=testtb-testConsecutiveExports in 618 msec 2025-01-08T05:00:04,157 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=81 2025-01-08T05:00:04,157 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testConsecutiveExports, procId: 81 completed 2025-01-08T05:00:04,160 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testConsecutiveExports 2025-01-08T05:00:04,160 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b. 2025-01-08T05:00:04,160 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T05:00:04,173 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } 2025-01-08T05:00:04,173 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736312404173 (current time:1736312404173). 2025-01-08T05:00:04,174 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T05:00:04,174 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot emptySnaptb0-testConsecutiveExports VERSION not specified, setting to 2 2025-01-08T05:00:04,174 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T05:00:04,175 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x6f536605 to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@43ada9f7 2025-01-08T05:00:04,191 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@6ae0753c, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:00:04,192 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:00:04,194 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:50310, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:00:04,195 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x6f536605 to 127.0.0.1:59048 2025-01-08T05:00:04,195 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:00:04,196 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x7c982e97 to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@5896082e 2025-01-08T05:00:04,223 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@3136df95, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:00:04,226 DEBUG [hconnection-0x1166129e-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:00:04,227 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:50326, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:00:04,233 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x7c982e97 to 127.0.0.1:59048 2025-01-08T05:00:04,233 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:00:04,234 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] access.PermissionStorage(611): Read acl: entry[testtb-testConsecutiveExports], kv [jenkins: RWXCA] 2025-01-08T05:00:04,235 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T05:00:04,239 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=86, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } 2025-01-08T05:00:04,239 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 }, snapshot procedure id = 86 2025-01-08T05:00:04,240 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T05:00:04,240 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=86 2025-01-08T05:00:04,241 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T05:00:04,245 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T05:00:04,261 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742038_1214 (size=161) 2025-01-08T05:00:04,261 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742038_1214 (size=161) 2025-01-08T05:00:04,262 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742038_1214 (size=161) 2025-01-08T05:00:04,269 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T05:00:04,269 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=87, ppid=86, state=RUNNABLE; SnapshotRegionProcedure b92a3e58f76a25816db26a2020c9df1b}, {pid=88, ppid=86, state=RUNNABLE; SnapshotRegionProcedure c432634a817b2f286fb158c1aad95aab}] 2025-01-08T05:00:04,271 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=88, ppid=86, state=RUNNABLE; SnapshotRegionProcedure c432634a817b2f286fb158c1aad95aab 2025-01-08T05:00:04,271 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=87, ppid=86, state=RUNNABLE; SnapshotRegionProcedure b92a3e58f76a25816db26a2020c9df1b 2025-01-08T05:00:04,341 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=86 2025-01-08T05:00:04,425 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:00:04,425 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,39801,1736312288779 2025-01-08T05:00:04,426 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=37967 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=87 2025-01-08T05:00:04,426 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=39801 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=88 2025-01-08T05:00:04,426 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=87}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b. 2025-01-08T05:00:04,426 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=88}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab. 2025-01-08T05:00:04,427 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=87}] regionserver.HRegion(2538): Flush status journal for b92a3e58f76a25816db26a2020c9df1b: 2025-01-08T05:00:04,427 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=88}] regionserver.HRegion(2538): Flush status journal for c432634a817b2f286fb158c1aad95aab: 2025-01-08T05:00:04,427 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=87}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b. for emptySnaptb0-testConsecutiveExports completed. 2025-01-08T05:00:04,427 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=88}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab. for emptySnaptb0-testConsecutiveExports completed. 2025-01-08T05:00:04,427 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=88}] snapshot.SnapshotManifest(241): Storing 'testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab.' region-info for snapshot=emptySnaptb0-testConsecutiveExports 2025-01-08T05:00:04,427 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=87}] snapshot.SnapshotManifest(241): Storing 'testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b.' region-info for snapshot=emptySnaptb0-testConsecutiveExports 2025-01-08T05:00:04,427 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=88}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T05:00:04,427 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=87}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T05:00:04,427 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=88}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T05:00:04,427 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=87}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T05:00:04,463 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742039_1215 (size=68) 2025-01-08T05:00:04,463 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742039_1215 (size=68) 2025-01-08T05:00:04,463 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742039_1215 (size=68) 2025-01-08T05:00:04,464 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=88}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab. 2025-01-08T05:00:04,464 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=88}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=88 2025-01-08T05:00:04,465 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=88 2025-01-08T05:00:04,465 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testConsecutiveExports on region c432634a817b2f286fb158c1aad95aab 2025-01-08T05:00:04,465 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=88, ppid=86, state=RUNNABLE; SnapshotRegionProcedure c432634a817b2f286fb158c1aad95aab 2025-01-08T05:00:04,467 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742040_1216 (size=68) 2025-01-08T05:00:04,468 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742040_1216 (size=68) 2025-01-08T05:00:04,468 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742040_1216 (size=68) 2025-01-08T05:00:04,468 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=88, ppid=86, state=SUCCESS; SnapshotRegionProcedure c432634a817b2f286fb158c1aad95aab in 198 msec 2025-01-08T05:00:04,469 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=87}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b. 2025-01-08T05:00:04,469 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=87}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=87 2025-01-08T05:00:04,470 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=87 2025-01-08T05:00:04,470 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testConsecutiveExports on region b92a3e58f76a25816db26a2020c9df1b 2025-01-08T05:00:04,470 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=87, ppid=86, state=RUNNABLE; SnapshotRegionProcedure b92a3e58f76a25816db26a2020c9df1b 2025-01-08T05:00:04,473 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=87, resume processing ppid=86 2025-01-08T05:00:04,473 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=87, ppid=86, state=SUCCESS; SnapshotRegionProcedure b92a3e58f76a25816db26a2020c9df1b in 202 msec 2025-01-08T05:00:04,473 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T05:00:04,475 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T05:00:04,484 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T05:00:04,484 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T05:00:04,484 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:00:04,485 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(328): No files under family: cf 2025-01-08T05:00:04,499 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742041_1217 (size=60) 2025-01-08T05:00:04,500 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742041_1217 (size=60) 2025-01-08T05:00:04,500 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742041_1217 (size=60) 2025-01-08T05:00:04,501 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T05:00:04,501 DEBUG [PEWorker-5 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for emptySnaptb0-testConsecutiveExports 2025-01-08T05:00:04,502 DEBUG [PEWorker-5 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/emptySnaptb0-testConsecutiveExports 2025-01-08T05:00:04,515 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742042_1218 (size=641) 2025-01-08T05:00:04,516 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742042_1218 (size=641) 2025-01-08T05:00:04,516 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742042_1218 (size=641) 2025-01-08T05:00:04,519 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T05:00:04,524 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T05:00:04,525 DEBUG [PEWorker-5 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/emptySnaptb0-testConsecutiveExports to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/emptySnaptb0-testConsecutiveExports 2025-01-08T05:00:04,529 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T05:00:04,529 DEBUG [PEWorker-5 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 }, snapshot procedure id = 86 2025-01-08T05:00:04,531 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=86, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } in 294 msec 2025-01-08T05:00:04,543 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=86 2025-01-08T05:00:04,543 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testConsecutiveExports, procId: 86 completed 2025-01-08T05:00:04,546 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37967 {}] regionserver.HRegion(8254): writing data to region testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T05:00:04,547 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39801 {}] regionserver.HRegion(8254): writing data to region testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T05:00:04,551 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testConsecutiveExports 2025-01-08T05:00:04,551 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b. 2025-01-08T05:00:04,551 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T05:00:04,565 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } 2025-01-08T05:00:04,565 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736312404565 (current time:1736312404565). 2025-01-08T05:00:04,565 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T05:00:04,565 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testConsecutiveExports VERSION not specified, setting to 2 2025-01-08T05:00:04,565 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T05:00:04,566 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x43457500 to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@62166d14 2025-01-08T05:00:04,583 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7563f107, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:00:04,585 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:00:04,586 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:35108, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:00:04,588 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x43457500 to 127.0.0.1:59048 2025-01-08T05:00:04,588 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:00:04,589 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x1f6fba08 to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@4e401a3c 2025-01-08T05:00:04,614 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2441213a, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:00:04,617 DEBUG [hconnection-0x2e9170e-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:00:04,618 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:35124, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:00:04,621 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x1f6fba08 to 127.0.0.1:59048 2025-01-08T05:00:04,621 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:00:04,621 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] access.PermissionStorage(611): Read acl: entry[testtb-testConsecutiveExports], kv [jenkins: RWXCA] 2025-01-08T05:00:04,622 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T05:00:04,623 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=89, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } 2025-01-08T05:00:04,623 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 }, snapshot procedure id = 89 2025-01-08T05:00:04,624 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T05:00:04,624 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=89 2025-01-08T05:00:04,625 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T05:00:04,627 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T05:00:04,670 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742043_1219 (size=156) 2025-01-08T05:00:04,671 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742043_1219 (size=156) 2025-01-08T05:00:04,671 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742043_1219 (size=156) 2025-01-08T05:00:04,726 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=89 2025-01-08T05:00:04,927 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=89 2025-01-08T05:00:05,065 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T05:00:05,065 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=90, ppid=89, state=RUNNABLE; SnapshotRegionProcedure b92a3e58f76a25816db26a2020c9df1b}, {pid=91, ppid=89, state=RUNNABLE; SnapshotRegionProcedure c432634a817b2f286fb158c1aad95aab}] 2025-01-08T05:00:05,066 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=91, ppid=89, state=RUNNABLE; SnapshotRegionProcedure c432634a817b2f286fb158c1aad95aab 2025-01-08T05:00:05,066 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=90, ppid=89, state=RUNNABLE; SnapshotRegionProcedure b92a3e58f76a25816db26a2020c9df1b 2025-01-08T05:00:05,218 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:00:05,218 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,39801,1736312288779 2025-01-08T05:00:05,218 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37967 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=90 2025-01-08T05:00:05,219 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b. 2025-01-08T05:00:05,219 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=39801 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=91 2025-01-08T05:00:05,219 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab. 2025-01-08T05:00:05,219 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.HRegion(2837): Flushing b92a3e58f76a25816db26a2020c9df1b 1/1 column families, dataSize=199 B heapSize=688 B 2025-01-08T05:00:05,219 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.HRegion(2837): Flushing c432634a817b2f286fb158c1aad95aab 1/1 column families, dataSize=3.06 KB heapSize=6.86 KB 2025-01-08T05:00:05,228 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=89 2025-01-08T05:00:05,257 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108e742e82edebb40e098ab082c8eaa98f6_b92a3e58f76a25816db26a2020c9df1b is 71, key is 0696741b92e50a0b9a25d2d7a66f6bc8/cf:q/1736312404546/Put/seqid=0 2025-01-08T05:00:05,267 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108146e4cca0ac44d36853c08c2f1cda3a4_c432634a817b2f286fb158c1aad95aab is 71, key is 126a28803b6b92fcf45731281cb15542/cf:q/1736312404547/Put/seqid=0 2025-01-08T05:00:05,297 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742045_1221 (size=8171) 2025-01-08T05:00:05,298 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742045_1221 (size=8171) 2025-01-08T05:00:05,298 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:00:05,298 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742045_1221 (size=8171) 2025-01-08T05:00:05,306 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108146e4cca0ac44d36853c08c2f1cda3a4_c432634a817b2f286fb158c1aad95aab to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/c4ca4238a0b923820dcc509a6f75849b20250108146e4cca0ac44d36853c08c2f1cda3a4_c432634a817b2f286fb158c1aad95aab 2025-01-08T05:00:05,311 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/c432634a817b2f286fb158c1aad95aab/.tmp/cf/eb892d1b008d47cf983a50572f4a907d, store: [table=testtb-testConsecutiveExports family=cf region=c432634a817b2f286fb158c1aad95aab] 2025-01-08T05:00:05,312 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/c432634a817b2f286fb158c1aad95aab/.tmp/cf/eb892d1b008d47cf983a50572f4a907d is 206, key is 19bda964af48e21a2eba05d2c7dee52ff/cf:q/1736312404547/Put/seqid=0 2025-01-08T05:00:05,326 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742044_1220 (size=5102) 2025-01-08T05:00:05,327 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742044_1220 (size=5102) 2025-01-08T05:00:05,327 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742044_1220 (size=5102) 2025-01-08T05:00:05,328 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:00:05,335 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108e742e82edebb40e098ab082c8eaa98f6_b92a3e58f76a25816db26a2020c9df1b to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/d41d8cd98f00b204e9800998ecf8427e20250108e742e82edebb40e098ab082c8eaa98f6_b92a3e58f76a25816db26a2020c9df1b 2025-01-08T05:00:05,337 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/b92a3e58f76a25816db26a2020c9df1b/.tmp/cf/4b8b2e91631943c797cc71f2db937e75, store: [table=testtb-testConsecutiveExports family=cf region=b92a3e58f76a25816db26a2020c9df1b] 2025-01-08T05:00:05,337 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/b92a3e58f76a25816db26a2020c9df1b/.tmp/cf/4b8b2e91631943c797cc71f2db937e75 is 206, key is 02c1542b9611225527371dfedac12e9c7/cf:q/1736312404546/Put/seqid=0 2025-01-08T05:00:05,340 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742046_1222 (size=14853) 2025-01-08T05:00:05,340 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742046_1222 (size=14853) 2025-01-08T05:00:05,340 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742046_1222 (size=14853) 2025-01-08T05:00:05,342 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=3.1 K, hasBloomFilter=true, into tmp file hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/c432634a817b2f286fb158c1aad95aab/.tmp/cf/eb892d1b008d47cf983a50572f4a907d 2025-01-08T05:00:05,352 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/c432634a817b2f286fb158c1aad95aab/.tmp/cf/eb892d1b008d47cf983a50572f4a907d as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/c432634a817b2f286fb158c1aad95aab/cf/eb892d1b008d47cf983a50572f4a907d 2025-01-08T05:00:05,359 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/c432634a817b2f286fb158c1aad95aab/cf/eb892d1b008d47cf983a50572f4a907d, entries=47, sequenceid=6, filesize=14.5 K 2025-01-08T05:00:05,360 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.HRegion(3040): Finished flush of dataSize ~3.06 KB/3137, heapSize ~6.84 KB/7008, currentSize=0 B/0 for c432634a817b2f286fb158c1aad95aab in 141ms, sequenceid=6, compaction requested=false 2025-01-08T05:00:05,360 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testConsecutiveExports' 2025-01-08T05:00:05,361 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.HRegion(2538): Flush status journal for c432634a817b2f286fb158c1aad95aab: 2025-01-08T05:00:05,361 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab. for snaptb0-testConsecutiveExports completed. 2025-01-08T05:00:05,361 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] snapshot.SnapshotManifest(241): Storing 'testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab.' region-info for snapshot=snaptb0-testConsecutiveExports 2025-01-08T05:00:05,361 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T05:00:05,362 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/c432634a817b2f286fb158c1aad95aab/cf/eb892d1b008d47cf983a50572f4a907d] hfiles 2025-01-08T05:00:05,362 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/c432634a817b2f286fb158c1aad95aab/cf/eb892d1b008d47cf983a50572f4a907d for snapshot=snaptb0-testConsecutiveExports 2025-01-08T05:00:05,393 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742047_1223 (size=5906) 2025-01-08T05:00:05,393 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742047_1223 (size=5906) 2025-01-08T05:00:05,394 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742047_1223 (size=5906) 2025-01-08T05:00:05,394 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=199, hasBloomFilter=true, into tmp file hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/b92a3e58f76a25816db26a2020c9df1b/.tmp/cf/4b8b2e91631943c797cc71f2db937e75 2025-01-08T05:00:05,403 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/b92a3e58f76a25816db26a2020c9df1b/.tmp/cf/4b8b2e91631943c797cc71f2db937e75 as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/b92a3e58f76a25816db26a2020c9df1b/cf/4b8b2e91631943c797cc71f2db937e75 2025-01-08T05:00:05,412 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/b92a3e58f76a25816db26a2020c9df1b/cf/4b8b2e91631943c797cc71f2db937e75, entries=3, sequenceid=6, filesize=5.8 K 2025-01-08T05:00:05,413 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.HRegion(3040): Finished flush of dataSize ~199 B/199, heapSize ~672 B/672, currentSize=0 B/0 for b92a3e58f76a25816db26a2020c9df1b in 194ms, sequenceid=6, compaction requested=false 2025-01-08T05:00:05,413 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.HRegion(2538): Flush status journal for b92a3e58f76a25816db26a2020c9df1b: 2025-01-08T05:00:05,414 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b. for snaptb0-testConsecutiveExports completed. 2025-01-08T05:00:05,414 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] snapshot.SnapshotManifest(241): Storing 'testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b.' region-info for snapshot=snaptb0-testConsecutiveExports 2025-01-08T05:00:05,414 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T05:00:05,414 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/b92a3e58f76a25816db26a2020c9df1b/cf/4b8b2e91631943c797cc71f2db937e75] hfiles 2025-01-08T05:00:05,414 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/b92a3e58f76a25816db26a2020c9df1b/cf/4b8b2e91631943c797cc71f2db937e75 for snapshot=snaptb0-testConsecutiveExports 2025-01-08T05:00:05,423 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742048_1224 (size=107) 2025-01-08T05:00:05,424 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742048_1224 (size=107) 2025-01-08T05:00:05,424 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742048_1224 (size=107) 2025-01-08T05:00:05,425 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab. 2025-01-08T05:00:05,425 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=91 2025-01-08T05:00:05,425 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=91 2025-01-08T05:00:05,425 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testConsecutiveExports on region c432634a817b2f286fb158c1aad95aab 2025-01-08T05:00:05,426 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=91, ppid=89, state=RUNNABLE; SnapshotRegionProcedure c432634a817b2f286fb158c1aad95aab 2025-01-08T05:00:05,431 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=91, ppid=89, state=SUCCESS; SnapshotRegionProcedure c432634a817b2f286fb158c1aad95aab in 362 msec 2025-01-08T05:00:05,444 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742049_1225 (size=107) 2025-01-08T05:00:05,444 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742049_1225 (size=107) 2025-01-08T05:00:05,446 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742049_1225 (size=107) 2025-01-08T05:00:05,446 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b. 2025-01-08T05:00:05,447 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=90 2025-01-08T05:00:05,451 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=90 2025-01-08T05:00:05,451 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testConsecutiveExports on region b92a3e58f76a25816db26a2020c9df1b 2025-01-08T05:00:05,451 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=90, ppid=89, state=RUNNABLE; SnapshotRegionProcedure b92a3e58f76a25816db26a2020c9df1b 2025-01-08T05:00:05,460 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=90, resume processing ppid=89 2025-01-08T05:00:05,460 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=90, ppid=89, state=SUCCESS; SnapshotRegionProcedure b92a3e58f76a25816db26a2020c9df1b in 390 msec 2025-01-08T05:00:05,460 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T05:00:05,462 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T05:00:05,464 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T05:00:05,464 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T05:00:05,464 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:00:05,466 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/c4ca4238a0b923820dcc509a6f75849b20250108146e4cca0ac44d36853c08c2f1cda3a4_c432634a817b2f286fb158c1aad95aab, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/d41d8cd98f00b204e9800998ecf8427e20250108e742e82edebb40e098ab082c8eaa98f6_b92a3e58f76a25816db26a2020c9df1b] hfiles 2025-01-08T05:00:05,466 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/c4ca4238a0b923820dcc509a6f75849b20250108146e4cca0ac44d36853c08c2f1cda3a4_c432634a817b2f286fb158c1aad95aab 2025-01-08T05:00:05,466 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/d41d8cd98f00b204e9800998ecf8427e20250108e742e82edebb40e098ab082c8eaa98f6_b92a3e58f76a25816db26a2020c9df1b 2025-01-08T05:00:05,480 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742050_1226 (size=291) 2025-01-08T05:00:05,480 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742050_1226 (size=291) 2025-01-08T05:00:05,481 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742050_1226 (size=291) 2025-01-08T05:00:05,483 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T05:00:05,483 DEBUG [PEWorker-2 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testConsecutiveExports 2025-01-08T05:00:05,484 DEBUG [PEWorker-2 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/snaptb0-testConsecutiveExports 2025-01-08T05:00:05,512 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742051_1227 (size=951) 2025-01-08T05:00:05,512 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742051_1227 (size=951) 2025-01-08T05:00:05,513 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742051_1227 (size=951) 2025-01-08T05:00:05,521 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T05:00:05,532 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T05:00:05,532 DEBUG [PEWorker-2 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/snaptb0-testConsecutiveExports to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testConsecutiveExports 2025-01-08T05:00:05,534 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T05:00:05,534 DEBUG [PEWorker-2 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 }, snapshot procedure id = 89 2025-01-08T05:00:05,535 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=89, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } in 912 msec 2025-01-08T05:00:05,730 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=89 2025-01-08T05:00:05,730 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testConsecutiveExports, procId: 89 completed 2025-01-08T05:00:05,730 INFO [Time-limited test {}] snapshot.TestExportSnapshot(476): Local export destination path: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312405730 2025-01-08T05:00:05,730 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=file:///, tgtDir=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312405730, rawTgtDir=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312405730, srcFsUri=hdfs://localhost:45939, srcDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:00:05,770 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:45939, inputRoot=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:00:05,770 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=org.apache.hadoop.fs.LocalFileSystem@2c860d76, outputRoot=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312405730, skipTmp=false, initialOutputSnapshotDir=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312405730/.hbase-snapshot/.tmp/snaptb0-testConsecutiveExports 2025-01-08T05:00:05,772 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T05:00:05,781 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testConsecutiveExports to file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312405730/.hbase-snapshot/.tmp/snaptb0-testConsecutiveExports 2025-01-08T05:00:05,790 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T05:00:05,877 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:05,877 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:05,877 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:05,878 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:06,661 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0003_000001 (auth:SIMPLE) from 127.0.0.1:57922 2025-01-08T05:00:07,302 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop-15931472959993106902.jar 2025-01-08T05:00:07,302 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:07,303 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:07,384 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop-8784710519783425220.jar 2025-01-08T05:00:07,385 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:07,385 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:07,385 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:07,386 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:07,386 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:07,386 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:07,387 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T05:00:07,387 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T05:00:07,388 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T05:00:07,388 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T05:00:07,388 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T05:00:07,388 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T05:00:07,389 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T05:00:07,389 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T05:00:07,389 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T05:00:07,390 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T05:00:07,390 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T05:00:07,391 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T05:00:07,391 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:00:07,391 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:00:07,392 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T05:00:07,392 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:00:07,392 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:00:07,393 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T05:00:07,393 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T05:00:07,563 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742052_1228 (size=127628) 2025-01-08T05:00:07,563 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742052_1228 (size=127628) 2025-01-08T05:00:07,566 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742052_1228 (size=127628) 2025-01-08T05:00:07,640 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742053_1229 (size=2172137) 2025-01-08T05:00:07,640 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742053_1229 (size=2172137) 2025-01-08T05:00:07,640 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742053_1229 (size=2172137) 2025-01-08T05:00:07,680 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742054_1230 (size=213228) 2025-01-08T05:00:07,687 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742054_1230 (size=213228) 2025-01-08T05:00:07,687 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742054_1230 (size=213228) 2025-01-08T05:00:07,724 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742055_1231 (size=1877034) 2025-01-08T05:00:07,727 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742055_1231 (size=1877034) 2025-01-08T05:00:07,727 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742055_1231 (size=1877034) 2025-01-08T05:00:07,746 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742056_1232 (size=533455) 2025-01-08T05:00:07,747 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742056_1232 (size=533455) 2025-01-08T05:00:07,747 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742056_1232 (size=533455) 2025-01-08T05:00:07,840 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742057_1233 (size=6350712) 2025-01-08T05:00:07,842 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742057_1233 (size=6350712) 2025-01-08T05:00:07,843 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742057_1233 (size=6350712) 2025-01-08T05:00:07,898 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742058_1234 (size=7280644) 2025-01-08T05:00:07,899 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742058_1234 (size=7280644) 2025-01-08T05:00:07,899 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742058_1234 (size=7280644) 2025-01-08T05:00:07,937 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742059_1235 (size=4188619) 2025-01-08T05:00:07,938 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742059_1235 (size=4188619) 2025-01-08T05:00:07,938 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742059_1235 (size=4188619) 2025-01-08T05:00:07,987 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742060_1236 (size=20406) 2025-01-08T05:00:07,987 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742060_1236 (size=20406) 2025-01-08T05:00:07,987 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742060_1236 (size=20406) 2025-01-08T05:00:08,051 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T05:00:08,345 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testConsecutiveExports 2025-01-08T05:00:08,345 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testConsecutiveExports Metrics about Tables on a single HBase RegionServer 2025-01-08T05:00:08,346 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemState 2025-01-08T05:00:08,400 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742061_1237 (size=75495) 2025-01-08T05:00:08,401 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742061_1237 (size=75495) 2025-01-08T05:00:08,401 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742061_1237 (size=75495) 2025-01-08T05:00:08,410 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742062_1238 (size=45609) 2025-01-08T05:00:08,410 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742062_1238 (size=45609) 2025-01-08T05:00:08,411 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742062_1238 (size=45609) 2025-01-08T05:00:08,419 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742063_1239 (size=451756) 2025-01-08T05:00:08,420 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742063_1239 (size=451756) 2025-01-08T05:00:08,420 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742063_1239 (size=451756) 2025-01-08T05:00:08,430 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742064_1240 (size=110084) 2025-01-08T05:00:08,431 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742064_1240 (size=110084) 2025-01-08T05:00:08,431 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742064_1240 (size=110084) 2025-01-08T05:00:08,442 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742065_1241 (size=1323991) 2025-01-08T05:00:08,442 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742065_1241 (size=1323991) 2025-01-08T05:00:08,442 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742065_1241 (size=1323991) 2025-01-08T05:00:08,459 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742066_1242 (size=23076) 2025-01-08T05:00:08,460 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742066_1242 (size=23076) 2025-01-08T05:00:08,460 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742066_1242 (size=23076) 2025-01-08T05:00:08,472 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742067_1243 (size=126803) 2025-01-08T05:00:08,473 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742067_1243 (size=126803) 2025-01-08T05:00:08,473 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742067_1243 (size=126803) 2025-01-08T05:00:08,492 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742068_1244 (size=322274) 2025-01-08T05:00:08,494 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742068_1244 (size=322274) 2025-01-08T05:00:08,494 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742068_1244 (size=322274) 2025-01-08T05:00:08,530 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742069_1245 (size=1832290) 2025-01-08T05:00:08,530 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742069_1245 (size=1832290) 2025-01-08T05:00:08,530 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742069_1245 (size=1832290) 2025-01-08T05:00:08,553 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742070_1246 (size=30081) 2025-01-08T05:00:08,553 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742070_1246 (size=30081) 2025-01-08T05:00:08,554 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742070_1246 (size=30081) 2025-01-08T05:00:08,576 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742071_1247 (size=53616) 2025-01-08T05:00:08,576 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742071_1247 (size=53616) 2025-01-08T05:00:08,577 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742071_1247 (size=53616) 2025-01-08T05:00:08,603 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742072_1248 (size=29229) 2025-01-08T05:00:08,603 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742072_1248 (size=29229) 2025-01-08T05:00:08,603 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742072_1248 (size=29229) 2025-01-08T05:00:09,083 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742073_1249 (size=169089) 2025-01-08T05:00:09,083 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742073_1249 (size=169089) 2025-01-08T05:00:09,083 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742073_1249 (size=169089) 2025-01-08T05:00:09,122 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742074_1250 (size=5175431) 2025-01-08T05:00:09,123 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742074_1250 (size=5175431) 2025-01-08T05:00:09,124 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742074_1250 (size=5175431) 2025-01-08T05:00:09,184 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742075_1251 (size=136454) 2025-01-08T05:00:09,184 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742075_1251 (size=136454) 2025-01-08T05:00:09,184 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742075_1251 (size=136454) 2025-01-08T05:00:09,238 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742076_1252 (size=907470) 2025-01-08T05:00:09,238 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742076_1252 (size=907470) 2025-01-08T05:00:09,238 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742076_1252 (size=907470) 2025-01-08T05:00:09,312 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742077_1253 (size=3317408) 2025-01-08T05:00:09,312 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742077_1253 (size=3317408) 2025-01-08T05:00:09,312 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742077_1253 (size=3317408) 2025-01-08T05:00:09,751 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742078_1254 (size=503880) 2025-01-08T05:00:09,751 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742078_1254 (size=503880) 2025-01-08T05:00:09,751 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742078_1254 (size=503880) 2025-01-08T05:00:10,200 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742079_1255 (size=4695811) 2025-01-08T05:00:10,201 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742079_1255 (size=4695811) 2025-01-08T05:00:10,201 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742079_1255 (size=4695811) 2025-01-08T05:00:10,203 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T05:00:10,207 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'snaptb0-testConsecutiveExports' hfile list 2025-01-08T05:00:10,215 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(751): export split=0 size=33.2 K 2025-01-08T05:00:10,239 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742080_1256 (size=714) 2025-01-08T05:00:10,239 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742080_1256 (size=714) 2025-01-08T05:00:10,243 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742080_1256 (size=714) 2025-01-08T05:00:10,268 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742081_1257 (size=15) 2025-01-08T05:00:10,269 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742081_1257 (size=15) 2025-01-08T05:00:10,269 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742081_1257 (size=15) 2025-01-08T05:00:10,292 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742082_1258 (size=304928) 2025-01-08T05:00:10,295 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742082_1258 (size=304928) 2025-01-08T05:00:10,295 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742082_1258 (size=304928) 2025-01-08T05:00:10,339 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T05:00:10,339 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T05:00:10,456 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0004_000001 (auth:SIMPLE) from 127.0.0.1:58624 2025-01-08T05:00:11,784 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-1_3/usercache/jenkins/appcache/application_1736312298603_0003/container_1736312298603_0003_01_000001/launch_container.sh] 2025-01-08T05:00:11,785 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-1_3/usercache/jenkins/appcache/application_1736312298603_0003/container_1736312298603_0003_01_000001/container_tokens] 2025-01-08T05:00:11,785 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-1_3/usercache/jenkins/appcache/application_1736312298603_0003/container_1736312298603_0003_01_000001/sysfs] 2025-01-08T05:00:13,849 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T05:00:15,524 DEBUG [master/f28094b72178:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region b92a3e58f76a25816db26a2020c9df1b changed from -1.0 to 0.0, refreshing cache 2025-01-08T05:00:15,524 DEBUG [master/f28094b72178:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region c432634a817b2f286fb158c1aad95aab changed from -1.0 to 0.0, refreshing cache 2025-01-08T05:00:19,101 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0004_000001 (auth:SIMPLE) from 127.0.0.1:52478 2025-01-08T05:00:19,494 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742083_1259 (size=350602) 2025-01-08T05:00:19,495 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742083_1259 (size=350602) 2025-01-08T05:00:19,495 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742083_1259 (size=350602) 2025-01-08T05:00:21,354 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0004_000001 (auth:SIMPLE) from 127.0.0.1:59852 2025-01-08T05:00:25,998 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742084_1260 (size=17447) 2025-01-08T05:00:25,999 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742084_1260 (size=17447) 2025-01-08T05:00:25,999 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742084_1260 (size=17447) 2025-01-08T05:00:26,016 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742085_1261 (size=462) 2025-01-08T05:00:26,017 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742085_1261 (size=462) 2025-01-08T05:00:26,018 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742085_1261 (size=462) 2025-01-08T05:00:26,062 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742086_1262 (size=17447) 2025-01-08T05:00:26,062 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742086_1262 (size=17447) 2025-01-08T05:00:26,062 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742086_1262 (size=17447) 2025-01-08T05:00:26,079 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_3/usercache/jenkins/appcache/application_1736312298603_0004/container_1736312298603_0004_01_000002/launch_container.sh] 2025-01-08T05:00:26,079 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_3/usercache/jenkins/appcache/application_1736312298603_0004/container_1736312298603_0004_01_000002/container_tokens] 2025-01-08T05:00:26,080 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_3/usercache/jenkins/appcache/application_1736312298603_0004/container_1736312298603_0004_01_000002/sysfs] 2025-01-08T05:00:26,133 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742087_1263 (size=350602) 2025-01-08T05:00:26,133 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742087_1263 (size=350602) 2025-01-08T05:00:26,134 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742087_1263 (size=350602) 2025-01-08T05:00:26,170 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0004_000001 (auth:SIMPLE) from 127.0.0.1:35960 2025-01-08T05:00:27,766 INFO [Time-limited test {}] snapshot.ExportSnapshot(1207): Finalize the Snapshot Export 2025-01-08T05:00:27,766 INFO [Time-limited test {}] snapshot.ExportSnapshot(1218): Verify the exported snapshot's expiration status and integrity. 2025-01-08T05:00:27,779 INFO [Time-limited test {}] snapshot.ExportSnapshot(1224): Export Completed: snaptb0-testConsecutiveExports 2025-01-08T05:00:27,779 INFO [Time-limited test {}] snapshot.TestExportSnapshot(362): Exported snapshot 2025-01-08T05:00:27,780 INFO [Time-limited test {}] snapshot.TestExportSnapshot(373): Verified filesystem state 2025-01-08T05:00:27,780 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1149249371_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testConsecutiveExports at hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testConsecutiveExports 2025-01-08T05:00:27,785 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testConsecutiveExports/.snapshotinfo 2025-01-08T05:00:27,785 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testConsecutiveExports/data.manifest 2025-01-08T05:00:27,785 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in org.apache.hadoop.fs.LocalFileSystem@2c860d76 in root file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312405730/.hbase-snapshot/snaptb0-testConsecutiveExports at file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312405730/.hbase-snapshot/snaptb0-testConsecutiveExports 2025-01-08T05:00:27,785 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312405730/.hbase-snapshot/snaptb0-testConsecutiveExports/data.manifest 2025-01-08T05:00:27,785 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312405730/.hbase-snapshot/snaptb0-testConsecutiveExports/.snapshotinfo 2025-01-08T05:00:27,788 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=file:///, tgtDir=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312405730, rawTgtDir=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312405730, srcFsUri=hdfs://localhost:45939, srcDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:00:27,826 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:45939, inputRoot=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:00:27,826 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=org.apache.hadoop.fs.LocalFileSystem@2c860d76, outputRoot=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312405730, skipTmp=false, initialOutputSnapshotDir=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312405730/.hbase-snapshot/.tmp/snaptb0-testConsecutiveExports 2025-01-08T05:00:27,834 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T05:00:27,845 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testConsecutiveExports to file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312405730/.hbase-snapshot/.tmp/snaptb0-testConsecutiveExports 2025-01-08T05:00:27,876 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:27,876 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:27,877 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:27,877 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:29,174 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop-17411111652075435359.jar 2025-01-08T05:00:29,175 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:29,175 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:29,256 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop-11480055448662863693.jar 2025-01-08T05:00:29,256 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:29,256 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:29,257 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:29,257 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:29,257 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:29,257 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:29,258 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T05:00:29,258 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T05:00:29,258 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T05:00:29,258 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T05:00:29,258 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T05:00:29,259 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T05:00:29,259 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T05:00:29,259 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T05:00:29,259 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T05:00:29,260 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T05:00:29,260 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T05:00:29,260 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T05:00:29,261 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:00:29,261 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:00:29,261 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T05:00:29,261 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:00:29,261 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:00:29,262 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T05:00:29,262 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T05:00:29,337 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742088_1264 (size=127628) 2025-01-08T05:00:29,338 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742088_1264 (size=127628) 2025-01-08T05:00:29,338 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742088_1264 (size=127628) 2025-01-08T05:00:29,364 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742089_1265 (size=2172137) 2025-01-08T05:00:29,365 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742089_1265 (size=2172137) 2025-01-08T05:00:29,366 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742089_1265 (size=2172137) 2025-01-08T05:00:29,384 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742090_1266 (size=213228) 2025-01-08T05:00:29,384 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742090_1266 (size=213228) 2025-01-08T05:00:29,384 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742090_1266 (size=213228) 2025-01-08T05:00:29,401 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742091_1267 (size=1877034) 2025-01-08T05:00:29,401 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742091_1267 (size=1877034) 2025-01-08T05:00:29,402 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742091_1267 (size=1877034) 2025-01-08T05:00:29,436 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742092_1268 (size=451756) 2025-01-08T05:00:29,436 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742092_1268 (size=451756) 2025-01-08T05:00:29,436 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742092_1268 (size=451756) 2025-01-08T05:00:29,466 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742093_1269 (size=533455) 2025-01-08T05:00:29,467 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742093_1269 (size=533455) 2025-01-08T05:00:29,470 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742093_1269 (size=533455) 2025-01-08T05:00:29,560 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742094_1270 (size=7280644) 2025-01-08T05:00:29,560 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742094_1270 (size=7280644) 2025-01-08T05:00:29,560 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742094_1270 (size=7280644) 2025-01-08T05:00:29,613 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742095_1271 (size=6350712) 2025-01-08T05:00:29,614 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742095_1271 (size=6350712) 2025-01-08T05:00:29,614 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742095_1271 (size=6350712) 2025-01-08T05:00:29,634 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742096_1272 (size=4188619) 2025-01-08T05:00:29,634 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742096_1272 (size=4188619) 2025-01-08T05:00:29,635 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742096_1272 (size=4188619) 2025-01-08T05:00:29,652 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742097_1273 (size=20406) 2025-01-08T05:00:29,653 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742097_1273 (size=20406) 2025-01-08T05:00:29,653 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742097_1273 (size=20406) 2025-01-08T05:00:29,682 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742098_1274 (size=75495) 2025-01-08T05:00:29,682 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742098_1274 (size=75495) 2025-01-08T05:00:29,683 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742098_1274 (size=75495) 2025-01-08T05:00:29,733 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742099_1275 (size=45609) 2025-01-08T05:00:29,733 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742099_1275 (size=45609) 2025-01-08T05:00:29,734 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742099_1275 (size=45609) 2025-01-08T05:00:30,187 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742100_1276 (size=110084) 2025-01-08T05:00:30,187 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742100_1276 (size=110084) 2025-01-08T05:00:30,188 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742100_1276 (size=110084) 2025-01-08T05:00:30,221 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742101_1277 (size=1323991) 2025-01-08T05:00:30,221 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742101_1277 (size=1323991) 2025-01-08T05:00:30,222 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742101_1277 (size=1323991) 2025-01-08T05:00:30,268 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742102_1278 (size=23076) 2025-01-08T05:00:30,269 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742102_1278 (size=23076) 2025-01-08T05:00:30,269 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742102_1278 (size=23076) 2025-01-08T05:00:30,292 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742103_1279 (size=126803) 2025-01-08T05:00:30,292 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742103_1279 (size=126803) 2025-01-08T05:00:30,293 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742103_1279 (size=126803) 2025-01-08T05:00:30,318 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742104_1280 (size=322274) 2025-01-08T05:00:30,318 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742104_1280 (size=322274) 2025-01-08T05:00:30,318 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742104_1280 (size=322274) 2025-01-08T05:00:30,795 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742105_1281 (size=1832290) 2025-01-08T05:00:30,799 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742105_1281 (size=1832290) 2025-01-08T05:00:30,799 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742105_1281 (size=1832290) 2025-01-08T05:00:30,827 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742106_1282 (size=30081) 2025-01-08T05:00:30,827 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742106_1282 (size=30081) 2025-01-08T05:00:30,827 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742106_1282 (size=30081) 2025-01-08T05:00:30,897 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742107_1283 (size=53616) 2025-01-08T05:00:30,897 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742107_1283 (size=53616) 2025-01-08T05:00:30,898 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742107_1283 (size=53616) 2025-01-08T05:00:30,927 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742108_1284 (size=29229) 2025-01-08T05:00:30,927 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742108_1284 (size=29229) 2025-01-08T05:00:30,928 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742108_1284 (size=29229) 2025-01-08T05:00:31,007 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742109_1285 (size=169089) 2025-01-08T05:00:31,007 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742109_1285 (size=169089) 2025-01-08T05:00:31,007 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742109_1285 (size=169089) 2025-01-08T05:00:31,559 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742110_1286 (size=5175431) 2025-01-08T05:00:31,559 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742110_1286 (size=5175431) 2025-01-08T05:00:31,559 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742110_1286 (size=5175431) 2025-01-08T05:00:31,665 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742111_1287 (size=136454) 2025-01-08T05:00:31,666 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742111_1287 (size=136454) 2025-01-08T05:00:31,667 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742111_1287 (size=136454) 2025-01-08T05:00:31,765 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742112_1288 (size=907470) 2025-01-08T05:00:31,768 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742112_1288 (size=907470) 2025-01-08T05:00:31,768 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742112_1288 (size=907470) 2025-01-08T05:00:31,885 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742113_1289 (size=3317408) 2025-01-08T05:00:31,886 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742113_1289 (size=3317408) 2025-01-08T05:00:31,887 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742113_1289 (size=3317408) 2025-01-08T05:00:31,968 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742114_1290 (size=503880) 2025-01-08T05:00:31,968 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742114_1290 (size=503880) 2025-01-08T05:00:31,974 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742114_1290 (size=503880) 2025-01-08T05:00:32,069 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742115_1291 (size=4695811) 2025-01-08T05:00:32,070 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742115_1291 (size=4695811) 2025-01-08T05:00:32,071 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742115_1291 (size=4695811) 2025-01-08T05:00:32,076 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T05:00:32,083 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'snaptb0-testConsecutiveExports' hfile list 2025-01-08T05:00:32,085 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(751): export split=0 size=33.2 K 2025-01-08T05:00:32,115 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742116_1292 (size=714) 2025-01-08T05:00:32,115 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742116_1292 (size=714) 2025-01-08T05:00:32,115 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742116_1292 (size=714) 2025-01-08T05:00:32,302 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0004_000001 (auth:SIMPLE) from 127.0.0.1:35972 2025-01-08T05:00:32,328 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_2/usercache/jenkins/appcache/application_1736312298603_0004/container_1736312298603_0004_01_000001/launch_container.sh] 2025-01-08T05:00:32,328 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_2/usercache/jenkins/appcache/application_1736312298603_0004/container_1736312298603_0004_01_000001/container_tokens] 2025-01-08T05:00:32,328 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_2/usercache/jenkins/appcache/application_1736312298603_0004/container_1736312298603_0004_01_000001/sysfs] 2025-01-08T05:00:32,565 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742117_1293 (size=15) 2025-01-08T05:00:32,566 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742117_1293 (size=15) 2025-01-08T05:00:32,566 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742117_1293 (size=15) 2025-01-08T05:00:32,662 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742118_1294 (size=304930) 2025-01-08T05:00:32,663 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742118_1294 (size=304930) 2025-01-08T05:00:32,663 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742118_1294 (size=304930) 2025-01-08T05:00:32,690 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T05:00:32,690 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T05:00:33,289 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0005_000001 (auth:SIMPLE) from 127.0.0.1:35986 2025-01-08T05:00:35,791 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T05:00:40,841 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0005_000001 (auth:SIMPLE) from 127.0.0.1:58762 2025-01-08T05:00:41,161 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742119_1295 (size=350604) 2025-01-08T05:00:41,179 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742119_1295 (size=350604) 2025-01-08T05:00:41,183 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742119_1295 (size=350604) 2025-01-08T05:00:43,309 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0005_000001 (auth:SIMPLE) from 127.0.0.1:57046 2025-01-08T05:00:48,496 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region b92a3e58f76a25816db26a2020c9df1b, had cached 0 bytes from a total of 5906 2025-01-08T05:00:48,505 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region c432634a817b2f286fb158c1aad95aab, had cached 0 bytes from a total of 14853 2025-01-08T05:00:48,895 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742120_1296 (size=16924) 2025-01-08T05:00:48,895 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742120_1296 (size=16924) 2025-01-08T05:00:48,896 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742120_1296 (size=16924) 2025-01-08T05:00:48,908 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742121_1297 (size=462) 2025-01-08T05:00:48,908 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742121_1297 (size=462) 2025-01-08T05:00:48,909 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742121_1297 (size=462) 2025-01-08T05:00:48,946 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_0/usercache/jenkins/appcache/application_1736312298603_0005/container_1736312298603_0005_01_000002/launch_container.sh] 2025-01-08T05:00:48,946 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_0/usercache/jenkins/appcache/application_1736312298603_0005/container_1736312298603_0005_01_000002/container_tokens] 2025-01-08T05:00:48,946 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_0/usercache/jenkins/appcache/application_1736312298603_0005/container_1736312298603_0005_01_000002/sysfs] 2025-01-08T05:00:48,947 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742122_1298 (size=16924) 2025-01-08T05:00:48,947 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742122_1298 (size=16924) 2025-01-08T05:00:48,947 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742122_1298 (size=16924) 2025-01-08T05:00:48,981 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742123_1299 (size=350604) 2025-01-08T05:00:48,981 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742123_1299 (size=350604) 2025-01-08T05:00:48,982 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742123_1299 (size=350604) 2025-01-08T05:00:50,919 INFO [Time-limited test {}] snapshot.ExportSnapshot(1207): Finalize the Snapshot Export 2025-01-08T05:00:50,919 INFO [Time-limited test {}] snapshot.ExportSnapshot(1218): Verify the exported snapshot's expiration status and integrity. 2025-01-08T05:00:50,928 INFO [Time-limited test {}] snapshot.ExportSnapshot(1224): Export Completed: snaptb0-testConsecutiveExports 2025-01-08T05:00:50,928 INFO [Time-limited test {}] snapshot.TestExportSnapshot(362): Exported snapshot 2025-01-08T05:00:50,928 INFO [Time-limited test {}] snapshot.TestExportSnapshot(373): Verified filesystem state 2025-01-08T05:00:50,928 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1149249371_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testConsecutiveExports at hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testConsecutiveExports 2025-01-08T05:00:50,931 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testConsecutiveExports/.snapshotinfo 2025-01-08T05:00:50,931 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testConsecutiveExports/data.manifest 2025-01-08T05:00:50,931 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in org.apache.hadoop.fs.LocalFileSystem@2c860d76 in root file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312405730/.hbase-snapshot/snaptb0-testConsecutiveExports at file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312405730/.hbase-snapshot/snaptb0-testConsecutiveExports 2025-01-08T05:00:50,931 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312405730/.hbase-snapshot/snaptb0-testConsecutiveExports/data.manifest 2025-01-08T05:00:50,931 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312405730/.hbase-snapshot/snaptb0-testConsecutiveExports/.snapshotinfo 2025-01-08T05:00:50,950 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testConsecutiveExports 2025-01-08T05:00:50,950 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.3 disable testtb-testConsecutiveExports 2025-01-08T05:00:50,952 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=92, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testConsecutiveExports 2025-01-08T05:00:50,955 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testConsecutiveExports","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312450954"}]},"ts":"1736312450954"} 2025-01-08T05:00:50,956 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testConsecutiveExports, state=DISABLING in hbase:meta 2025-01-08T05:00:50,958 INFO [PEWorker-1 {}] procedure.DisableTableProcedure(284): Set testtb-testConsecutiveExports to state=DISABLING 2025-01-08T05:00:50,959 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=92 2025-01-08T05:00:50,959 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=93, ppid=92, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testConsecutiveExports}] 2025-01-08T05:00:50,961 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=94, ppid=93, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=b92a3e58f76a25816db26a2020c9df1b, UNASSIGN}, {pid=95, ppid=93, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=c432634a817b2f286fb158c1aad95aab, UNASSIGN}] 2025-01-08T05:00:50,961 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=95, ppid=93, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=c432634a817b2f286fb158c1aad95aab, UNASSIGN 2025-01-08T05:00:50,962 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=94, ppid=93, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=b92a3e58f76a25816db26a2020c9df1b, UNASSIGN 2025-01-08T05:00:50,962 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=95 updating hbase:meta row=c432634a817b2f286fb158c1aad95aab, regionState=CLOSING, regionLocation=f28094b72178,39801,1736312288779 2025-01-08T05:00:50,963 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=94 updating hbase:meta row=b92a3e58f76a25816db26a2020c9df1b, regionState=CLOSING, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T05:00:50,965 DEBUG [PEWorker-3 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T05:00:50,965 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=96, ppid=95, state=RUNNABLE; CloseRegionProcedure c432634a817b2f286fb158c1aad95aab, server=f28094b72178,39801,1736312288779}] 2025-01-08T05:00:50,966 DEBUG [PEWorker-5 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T05:00:50,966 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=97, ppid=94, state=RUNNABLE; CloseRegionProcedure b92a3e58f76a25816db26a2020c9df1b, server=f28094b72178,37967,1736312288941}] 2025-01-08T05:00:51,060 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=92 2025-01-08T05:00:51,117 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,39801,1736312288779 2025-01-08T05:00:51,118 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] handler.UnassignRegionHandler(124): Close c432634a817b2f286fb158c1aad95aab 2025-01-08T05:00:51,118 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:00:51,118 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T05:00:51,119 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] regionserver.HRegion(1681): Closing c432634a817b2f286fb158c1aad95aab, disabling compactions & flushes 2025-01-08T05:00:51,119 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] regionserver.HRegion(1703): Closing region testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab. 2025-01-08T05:00:51,119 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab. 2025-01-08T05:00:51,119 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] regionserver.HRegion(1791): Acquired close lock on testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab. after waiting 0 ms 2025-01-08T05:00:51,119 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] regionserver.HRegion(1801): Updates disabled for region testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab. 2025-01-08T05:00:51,119 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] handler.UnassignRegionHandler(124): Close b92a3e58f76a25816db26a2020c9df1b 2025-01-08T05:00:51,119 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T05:00:51,119 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] regionserver.HRegion(1681): Closing b92a3e58f76a25816db26a2020c9df1b, disabling compactions & flushes 2025-01-08T05:00:51,119 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] regionserver.HRegion(1703): Closing region testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b. 2025-01-08T05:00:51,119 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b. 2025-01-08T05:00:51,119 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] regionserver.HRegion(1791): Acquired close lock on testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b. after waiting 0 ms 2025-01-08T05:00:51,119 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] regionserver.HRegion(1801): Updates disabled for region testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b. 2025-01-08T05:00:51,123 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/b92a3e58f76a25816db26a2020c9df1b/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T05:00:51,124 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:00:51,124 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] regionserver.HRegion(1922): Closed testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b. 2025-01-08T05:00:51,124 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] regionserver.HRegion(1635): Region close journal for b92a3e58f76a25816db26a2020c9df1b: 2025-01-08T05:00:51,127 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/c432634a817b2f286fb158c1aad95aab/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T05:00:51,127 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:00:51,127 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] regionserver.HRegion(1922): Closed testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab. 2025-01-08T05:00:51,127 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] regionserver.HRegion(1635): Region close journal for c432634a817b2f286fb158c1aad95aab: 2025-01-08T05:00:51,128 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] handler.UnassignRegionHandler(170): Closed b92a3e58f76a25816db26a2020c9df1b 2025-01-08T05:00:51,131 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=94 updating hbase:meta row=b92a3e58f76a25816db26a2020c9df1b, regionState=CLOSED 2025-01-08T05:00:51,131 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] handler.UnassignRegionHandler(170): Closed c432634a817b2f286fb158c1aad95aab 2025-01-08T05:00:51,132 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=95 updating hbase:meta row=c432634a817b2f286fb158c1aad95aab, regionState=CLOSED 2025-01-08T05:00:51,136 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=97, resume processing ppid=94 2025-01-08T05:00:51,136 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=96, resume processing ppid=95 2025-01-08T05:00:51,136 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=96, ppid=95, state=SUCCESS; CloseRegionProcedure c432634a817b2f286fb158c1aad95aab, server=f28094b72178,39801,1736312288779 in 168 msec 2025-01-08T05:00:51,137 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=97, ppid=94, state=SUCCESS; CloseRegionProcedure b92a3e58f76a25816db26a2020c9df1b, server=f28094b72178,37967,1736312288941 in 166 msec 2025-01-08T05:00:51,137 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=94, ppid=93, state=SUCCESS; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=b92a3e58f76a25816db26a2020c9df1b, UNASSIGN in 175 msec 2025-01-08T05:00:51,138 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=95, resume processing ppid=93 2025-01-08T05:00:51,138 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=95, ppid=93, state=SUCCESS; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=c432634a817b2f286fb158c1aad95aab, UNASSIGN in 175 msec 2025-01-08T05:00:51,140 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=93, resume processing ppid=92 2025-01-08T05:00:51,140 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=93, ppid=92, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testConsecutiveExports in 179 msec 2025-01-08T05:00:51,142 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testConsecutiveExports","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312451142"}]},"ts":"1736312451142"} 2025-01-08T05:00:51,144 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testConsecutiveExports, state=DISABLED in hbase:meta 2025-01-08T05:00:51,146 INFO [PEWorker-2 {}] procedure.DisableTableProcedure(296): Set testtb-testConsecutiveExports to state=DISABLED 2025-01-08T05:00:51,147 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=92, state=SUCCESS; DisableTableProcedure table=testtb-testConsecutiveExports in 196 msec 2025-01-08T05:00:51,261 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=92 2025-01-08T05:00:51,261 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testConsecutiveExports, procId: 92 completed 2025-01-08T05:00:51,262 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.3 delete testtb-testConsecutiveExports 2025-01-08T05:00:51,263 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=98, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testConsecutiveExports 2025-01-08T05:00:51,264 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=98, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testConsecutiveExports 2025-01-08T05:00:51,264 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testConsecutiveExports 2025-01-08T05:00:51,265 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=98, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testConsecutiveExports 2025-01-08T05:00:51,266 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37583 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testConsecutiveExports 2025-01-08T05:00:51,268 DEBUG [HFileArchiver-9 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/b92a3e58f76a25816db26a2020c9df1b 2025-01-08T05:00:51,268 DEBUG [HFileArchiver-10 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/c432634a817b2f286fb158c1aad95aab 2025-01-08T05:00:51,269 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testConsecutiveExports 2025-01-08T05:00:51,269 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testConsecutiveExports 2025-01-08T05:00:51,269 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testConsecutiveExports 2025-01-08T05:00:51,270 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testConsecutiveExports 2025-01-08T05:00:51,270 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testConsecutiveExports with data PBUF 2025-01-08T05:00:51,270 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testConsecutiveExports with data PBUF 2025-01-08T05:00:51,270 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testConsecutiveExports with data PBUF 2025-01-08T05:00:51,270 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testConsecutiveExports with data PBUF 2025-01-08T05:00:51,271 DEBUG [HFileArchiver-9 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/b92a3e58f76a25816db26a2020c9df1b/cf, FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/b92a3e58f76a25816db26a2020c9df1b/recovered.edits] 2025-01-08T05:00:51,271 DEBUG [HFileArchiver-10 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/c432634a817b2f286fb158c1aad95aab/cf, FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/c432634a817b2f286fb158c1aad95aab/recovered.edits] 2025-01-08T05:00:51,272 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testConsecutiveExports 2025-01-08T05:00:51,272 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testConsecutiveExports 2025-01-08T05:00:51,272 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:00:51,272 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:00:51,272 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testConsecutiveExports 2025-01-08T05:00:51,272 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testConsecutiveExports 2025-01-08T05:00:51,272 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:00:51,272 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:00:51,273 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=98 2025-01-08T05:00:51,275 DEBUG [HFileArchiver-10 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/c432634a817b2f286fb158c1aad95aab/cf/eb892d1b008d47cf983a50572f4a907d to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testConsecutiveExports/c432634a817b2f286fb158c1aad95aab/cf/eb892d1b008d47cf983a50572f4a907d 2025-01-08T05:00:51,276 DEBUG [HFileArchiver-9 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/b92a3e58f76a25816db26a2020c9df1b/cf/4b8b2e91631943c797cc71f2db937e75 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testConsecutiveExports/b92a3e58f76a25816db26a2020c9df1b/cf/4b8b2e91631943c797cc71f2db937e75 2025-01-08T05:00:51,279 DEBUG [HFileArchiver-9 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/b92a3e58f76a25816db26a2020c9df1b/recovered.edits/9.seqid to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testConsecutiveExports/b92a3e58f76a25816db26a2020c9df1b/recovered.edits/9.seqid 2025-01-08T05:00:51,279 DEBUG [HFileArchiver-10 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/c432634a817b2f286fb158c1aad95aab/recovered.edits/9.seqid to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testConsecutiveExports/c432634a817b2f286fb158c1aad95aab/recovered.edits/9.seqid 2025-01-08T05:00:51,279 DEBUG [HFileArchiver-9 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/b92a3e58f76a25816db26a2020c9df1b 2025-01-08T05:00:51,279 DEBUG [HFileArchiver-10 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testConsecutiveExports/c432634a817b2f286fb158c1aad95aab 2025-01-08T05:00:51,279 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(313): Archived testtb-testConsecutiveExports regions 2025-01-08T05:00:51,280 DEBUG [PEWorker-1 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac 2025-01-08T05:00:51,281 DEBUG [PEWorker-1 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf] 2025-01-08T05:00:51,284 DEBUG [PEWorker-1 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/c4ca4238a0b923820dcc509a6f75849b20250108146e4cca0ac44d36853c08c2f1cda3a4_c432634a817b2f286fb158c1aad95aab to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/c4ca4238a0b923820dcc509a6f75849b20250108146e4cca0ac44d36853c08c2f1cda3a4_c432634a817b2f286fb158c1aad95aab 2025-01-08T05:00:51,286 DEBUG [PEWorker-1 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/d41d8cd98f00b204e9800998ecf8427e20250108e742e82edebb40e098ab082c8eaa98f6_b92a3e58f76a25816db26a2020c9df1b to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/d41d8cd98f00b204e9800998ecf8427e20250108e742e82edebb40e098ab082c8eaa98f6_b92a3e58f76a25816db26a2020c9df1b 2025-01-08T05:00:51,286 DEBUG [PEWorker-1 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac 2025-01-08T05:00:51,288 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=98, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testConsecutiveExports 2025-01-08T05:00:51,291 WARN [PEWorker-1 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testtb-testConsecutiveExports from hbase:meta 2025-01-08T05:00:51,293 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testConsecutiveExports' descriptor. 2025-01-08T05:00:51,295 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=98, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testConsecutiveExports 2025-01-08T05:00:51,295 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testConsecutiveExports' from region states. 2025-01-08T05:00:51,295 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736312451295"}]},"ts":"9223372036854775807"} 2025-01-08T05:00:51,295 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736312451295"}]},"ts":"9223372036854775807"} 2025-01-08T05:00:51,297 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T05:00:51,297 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => b92a3e58f76a25816db26a2020c9df1b, NAME => 'testtb-testConsecutiveExports,,1736312403040.b92a3e58f76a25816db26a2020c9df1b.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => c432634a817b2f286fb158c1aad95aab, NAME => 'testtb-testConsecutiveExports,1,1736312403040.c432634a817b2f286fb158c1aad95aab.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T05:00:51,297 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testConsecutiveExports' as deleted. 2025-01-08T05:00:51,297 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testConsecutiveExports","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736312451297"}]},"ts":"9223372036854775807"} 2025-01-08T05:00:51,299 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testConsecutiveExports state from META 2025-01-08T05:00:51,301 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(133): Finished pid=98, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testConsecutiveExports 2025-01-08T05:00:51,302 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=98, state=SUCCESS; DeleteTableProcedure table=testtb-testConsecutiveExports in 39 msec 2025-01-08T05:00:51,374 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=98 2025-01-08T05:00:51,374 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testConsecutiveExports, procId: 98 completed 2025-01-08T05:00:51,384 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "emptySnaptb0-testConsecutiveExports" 2025-01-08T05:00:51,385 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotManager(380): Deleting snapshot: emptySnaptb0-testConsecutiveExports 2025-01-08T05:00:51,387 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "snaptb0-testConsecutiveExports" 2025-01-08T05:00:51,389 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testConsecutiveExports 2025-01-08T05:00:51,415 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: snapshot.TestMobSecureExportSnapshot#testConsecutiveExports Thread=787 (was 786) Potentially hanging thread: HFileArchiver-9 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #5 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #6 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_COMPACTED_FILES_DISCHARGER-regionserver/f28094b72178:0-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x168697cd-shared-pool-23 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (366384994) connection to localhost/127.0.0.1:35787 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: hconnection-0x168697cd-shared-pool-22 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1822135013_1 at /127.0.0.1:51198 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #8 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x168697cd-shared-pool-25 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x168697cd-shared-pool-24 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1149249371_22 at /127.0.0.1:48102 [Waiting for operation #5] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LogDeleter #2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1177) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_COMPACTED_FILES_DISCHARGER-regionserver/f28094b72178:0-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: process reaper (pid 18658) java.base@17.0.11/java.lang.ProcessHandleImpl.waitForProcessExit0(Native Method) java.base@17.0.11/java.lang.ProcessHandleImpl$1.run(ProcessHandleImpl.java:150) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Thread-4215 java.base@17.0.11/java.io.FileInputStream.readBytes(Native Method) java.base@17.0.11/java.io.FileInputStream.read(FileInputStream.java:276) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:282) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) java.base@17.0.11/sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:281) java.base@17.0.11/sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:324) java.base@17.0.11/sun.nio.cs.StreamDecoder.read(StreamDecoder.java:189) java.base@17.0.11/java.io.InputStreamReader.read(InputStreamReader.java:177) java.base@17.0.11/java.io.BufferedReader.fill(BufferedReader.java:162) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:329) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:396) app//org.apache.hadoop.util.Shell$1.run(Shell.java:1025) Potentially hanging thread: ApplicationMasterLauncher #7 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1822135013_1 at /127.0.0.1:35678 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-10 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:35787 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_COMPACTED_FILES_DISCHARGER-regionserver/f28094b72178:0-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LogDeleter #3 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1177) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) - Thread LEAK? -, OpenFileDescriptor=799 (was 803), MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=1164 (was 1141) - SystemLoadAverage LEAK? -, ProcessCount=17 (was 17), AvailableMemoryMB=2724 (was 3993) 2025-01-08T05:00:51,415 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=787 is superior to 500 2025-01-08T05:00:51,437 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: snapshot.TestMobSecureExportSnapshot#testExportFileSystemStateWithMergeRegion Thread=787, OpenFileDescriptor=799, MaxFileDescriptor=1048576, SystemLoadAverage=1164, ProcessCount=17, AvailableMemoryMB=2712 2025-01-08T05:00:51,437 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=787 is superior to 500 2025-01-08T05:00:51,439 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.3 create 'testtb-testExportFileSystemStateWithMergeRegion', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T05:00:51,440 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=99, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T05:00:51,442 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=99, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T05:00:51,443 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "testtb-testExportFileSystemStateWithMergeRegion" procId is: 99 2025-01-08T05:00:51,444 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=99 2025-01-08T05:00:51,444 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=99, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T05:00:51,464 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742124_1300 (size=458) 2025-01-08T05:00:51,465 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742124_1300 (size=458) 2025-01-08T05:00:51,465 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742124_1300 (size=458) 2025-01-08T05:00:51,468 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 762016803bad87dc082747011636b848, NAME => 'testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testtb-testExportFileSystemStateWithMergeRegion', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:00:51,468 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => 8ceb1f1fe1c86dfb459738a7200e1333, NAME => 'testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testtb-testExportFileSystemStateWithMergeRegion', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:00:51,486 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742125_1301 (size=83) 2025-01-08T05:00:51,486 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742125_1301 (size=83) 2025-01-08T05:00:51,486 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742125_1301 (size=83) 2025-01-08T05:00:51,487 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:00:51,488 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-0 {}] regionserver.HRegion(1681): Closing 762016803bad87dc082747011636b848, disabling compactions & flushes 2025-01-08T05:00:51,488 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848. 2025-01-08T05:00:51,488 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848. 2025-01-08T05:00:51,488 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848. after waiting 0 ms 2025-01-08T05:00:51,488 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848. 2025-01-08T05:00:51,488 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848. 2025-01-08T05:00:51,488 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-0 {}] regionserver.HRegion(1635): Region close journal for 762016803bad87dc082747011636b848: 2025-01-08T05:00:51,492 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742126_1302 (size=83) 2025-01-08T05:00:51,492 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742126_1302 (size=83) 2025-01-08T05:00:51,492 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742126_1302 (size=83) 2025-01-08T05:00:51,493 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:00:51,493 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-1 {}] regionserver.HRegion(1681): Closing 8ceb1f1fe1c86dfb459738a7200e1333, disabling compactions & flushes 2025-01-08T05:00:51,493 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333. 2025-01-08T05:00:51,493 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333. 2025-01-08T05:00:51,493 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333. after waiting 0 ms 2025-01-08T05:00:51,493 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333. 2025-01-08T05:00:51,493 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333. 2025-01-08T05:00:51,493 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-1 {}] regionserver.HRegion(1635): Region close journal for 8ceb1f1fe1c86dfb459738a7200e1333: 2025-01-08T05:00:51,494 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=99, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T05:00:51,495 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848.","families":{"info":[{"qualifier":"regioninfo","vlen":82,"tag":[],"timestamp":"1736312451495"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736312451495"}]},"ts":"1736312451495"} 2025-01-08T05:00:51,495 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333.","families":{"info":[{"qualifier":"regioninfo","vlen":82,"tag":[],"timestamp":"1736312451495"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736312451495"}]},"ts":"1736312451495"} 2025-01-08T05:00:51,500 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T05:00:51,501 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=99, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T05:00:51,501 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312451501"}]},"ts":"1736312451501"} 2025-01-08T05:00:51,502 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithMergeRegion, state=ENABLING in hbase:meta 2025-01-08T05:00:51,506 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(202): Hosts are {f28094b72178=0} racks are {/default-rack=0} 2025-01-08T05:00:51,508 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T05:00:51,508 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T05:00:51,508 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T05:00:51,508 INFO [PEWorker-4 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T05:00:51,508 INFO [PEWorker-4 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T05:00:51,508 INFO [PEWorker-4 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T05:00:51,508 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T05:00:51,508 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=100, ppid=99, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=762016803bad87dc082747011636b848, ASSIGN}, {pid=101, ppid=99, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=8ceb1f1fe1c86dfb459738a7200e1333, ASSIGN}] 2025-01-08T05:00:51,509 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=101, ppid=99, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=8ceb1f1fe1c86dfb459738a7200e1333, ASSIGN 2025-01-08T05:00:51,509 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=100, ppid=99, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=762016803bad87dc082747011636b848, ASSIGN 2025-01-08T05:00:51,510 INFO [PEWorker-5 {}] assignment.TransitRegionStateProcedure(264): Starting pid=100, ppid=99, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=762016803bad87dc082747011636b848, ASSIGN; state=OFFLINE, location=f28094b72178,37967,1736312288941; forceNewPlan=false, retain=false 2025-01-08T05:00:51,510 INFO [PEWorker-3 {}] assignment.TransitRegionStateProcedure(264): Starting pid=101, ppid=99, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=8ceb1f1fe1c86dfb459738a7200e1333, ASSIGN; state=OFFLINE, location=f28094b72178,37583,1736312288873; forceNewPlan=false, retain=false 2025-01-08T05:00:51,549 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=99 2025-01-08T05:00:51,660 INFO [f28094b72178:42195 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T05:00:51,660 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=101 updating hbase:meta row=8ceb1f1fe1c86dfb459738a7200e1333, regionState=OPENING, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T05:00:51,660 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=100 updating hbase:meta row=762016803bad87dc082747011636b848, regionState=OPENING, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T05:00:51,662 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=102, ppid=101, state=RUNNABLE; OpenRegionProcedure 8ceb1f1fe1c86dfb459738a7200e1333, server=f28094b72178,37583,1736312288873}] 2025-01-08T05:00:51,663 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=103, ppid=100, state=RUNNABLE; OpenRegionProcedure 762016803bad87dc082747011636b848, server=f28094b72178,37967,1736312288941}] 2025-01-08T05:00:51,750 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=99 2025-01-08T05:00:51,814 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T05:00:51,815 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:00:51,818 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] handler.AssignRegionHandler(135): Open testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848. 2025-01-08T05:00:51,818 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] handler.AssignRegionHandler(135): Open testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333. 2025-01-08T05:00:51,819 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(7285): Opening region: {ENCODED => 8ceb1f1fe1c86dfb459738a7200e1333, NAME => 'testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T05:00:51,819 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(7285): Opening region: {ENCODED => 762016803bad87dc082747011636b848, NAME => 'testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T05:00:51,819 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848. service=AccessControlService 2025-01-08T05:00:51,819 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333. service=AccessControlService 2025-01-08T05:00:51,819 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T05:00:51,819 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T05:00:51,819 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportFileSystemStateWithMergeRegion 762016803bad87dc082747011636b848 2025-01-08T05:00:51,819 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportFileSystemStateWithMergeRegion 8ceb1f1fe1c86dfb459738a7200e1333 2025-01-08T05:00:51,819 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:00:51,819 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:00:51,819 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(7327): checking encryption for 8ceb1f1fe1c86dfb459738a7200e1333 2025-01-08T05:00:51,819 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(7327): checking encryption for 762016803bad87dc082747011636b848 2025-01-08T05:00:51,819 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(7330): checking classloading for 8ceb1f1fe1c86dfb459738a7200e1333 2025-01-08T05:00:51,819 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(7330): checking classloading for 762016803bad87dc082747011636b848 2025-01-08T05:00:51,821 INFO [StoreOpener-762016803bad87dc082747011636b848-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 762016803bad87dc082747011636b848 2025-01-08T05:00:51,821 INFO [StoreOpener-8ceb1f1fe1c86dfb459738a7200e1333-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 8ceb1f1fe1c86dfb459738a7200e1333 2025-01-08T05:00:51,823 INFO [StoreOpener-762016803bad87dc082747011636b848-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 762016803bad87dc082747011636b848 columnFamilyName cf 2025-01-08T05:00:51,823 INFO [StoreOpener-8ceb1f1fe1c86dfb459738a7200e1333-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 8ceb1f1fe1c86dfb459738a7200e1333 columnFamilyName cf 2025-01-08T05:00:51,824 DEBUG [StoreOpener-762016803bad87dc082747011636b848-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:00:51,824 DEBUG [StoreOpener-8ceb1f1fe1c86dfb459738a7200e1333-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:00:51,824 INFO [StoreOpener-8ceb1f1fe1c86dfb459738a7200e1333-1 {}] regionserver.HStore(327): Store=8ceb1f1fe1c86dfb459738a7200e1333/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T05:00:51,824 INFO [StoreOpener-762016803bad87dc082747011636b848-1 {}] regionserver.HStore(327): Store=762016803bad87dc082747011636b848/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T05:00:51,825 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/8ceb1f1fe1c86dfb459738a7200e1333 2025-01-08T05:00:51,825 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/762016803bad87dc082747011636b848 2025-01-08T05:00:51,826 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/8ceb1f1fe1c86dfb459738a7200e1333 2025-01-08T05:00:51,826 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/762016803bad87dc082747011636b848 2025-01-08T05:00:51,828 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(1085): writing seq id for 762016803bad87dc082747011636b848 2025-01-08T05:00:51,828 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(1085): writing seq id for 8ceb1f1fe1c86dfb459738a7200e1333 2025-01-08T05:00:51,830 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/762016803bad87dc082747011636b848/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T05:00:51,830 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/8ceb1f1fe1c86dfb459738a7200e1333/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T05:00:51,830 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(1102): Opened 762016803bad87dc082747011636b848; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=62879147, jitterRate=-0.06302769482135773}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T05:00:51,830 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(1102): Opened 8ceb1f1fe1c86dfb459738a7200e1333; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=58811283, jitterRate=-0.12364359200000763}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T05:00:51,831 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(1001): Region open journal for 762016803bad87dc082747011636b848: 2025-01-08T05:00:51,831 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(1001): Region open journal for 8ceb1f1fe1c86dfb459738a7200e1333: 2025-01-08T05:00:51,832 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848., pid=103, masterSystemTime=1736312451815 2025-01-08T05:00:51,832 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333., pid=102, masterSystemTime=1736312451814 2025-01-08T05:00:51,833 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848. 2025-01-08T05:00:51,833 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] handler.AssignRegionHandler(164): Opened testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848. 2025-01-08T05:00:51,834 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=100 updating hbase:meta row=762016803bad87dc082747011636b848, regionState=OPEN, openSeqNum=2, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T05:00:51,834 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333. 2025-01-08T05:00:51,834 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] handler.AssignRegionHandler(164): Opened testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333. 2025-01-08T05:00:51,834 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=101 updating hbase:meta row=8ceb1f1fe1c86dfb459738a7200e1333, regionState=OPEN, openSeqNum=2, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T05:00:51,837 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=103, resume processing ppid=100 2025-01-08T05:00:51,837 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=103, ppid=100, state=SUCCESS; OpenRegionProcedure 762016803bad87dc082747011636b848, server=f28094b72178,37967,1736312288941 in 172 msec 2025-01-08T05:00:51,838 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=102, resume processing ppid=101 2025-01-08T05:00:51,838 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=102, ppid=101, state=SUCCESS; OpenRegionProcedure 8ceb1f1fe1c86dfb459738a7200e1333, server=f28094b72178,37583,1736312288873 in 174 msec 2025-01-08T05:00:51,838 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=100, ppid=99, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=762016803bad87dc082747011636b848, ASSIGN in 329 msec 2025-01-08T05:00:51,839 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=101, resume processing ppid=99 2025-01-08T05:00:51,839 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=101, ppid=99, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=8ceb1f1fe1c86dfb459738a7200e1333, ASSIGN in 330 msec 2025-01-08T05:00:51,840 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=99, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T05:00:51,840 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312451840"}]},"ts":"1736312451840"} 2025-01-08T05:00:51,841 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithMergeRegion, state=ENABLED in hbase:meta 2025-01-08T05:00:51,849 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=99, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T05:00:51,849 DEBUG [PEWorker-5 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testExportFileSystemStateWithMergeRegion jenkins: RWXCA 2025-01-08T05:00:51,851 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37583 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemStateWithMergeRegion], kv [jenkins: RWXCA] 2025-01-08T05:00:51,853 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:00:51,853 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:00:51,853 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:00:51,853 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:00:51,855 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T05:00:51,855 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T05:00:51,855 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T05:00:51,856 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T05:00:51,857 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=99, state=SUCCESS; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion in 416 msec 2025-01-08T05:00:52,051 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=99 2025-01-08T05:00:52,051 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion, procId: 99 completed 2025-01-08T05:00:52,054 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T05:00:52,054 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848. 2025-01-08T05:00:52,054 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T05:00:52,066 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } 2025-01-08T05:00:52,066 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736312452066 (current time:1736312452066). 2025-01-08T05:00:52,066 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T05:00:52,066 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot emptySnaptb0-testExportFileSystemStateWithMergeRegion VERSION not specified, setting to 2 2025-01-08T05:00:52,066 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T05:00:52,067 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x3257e4ca to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@161a891e 2025-01-08T05:00:52,071 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1c06c4bc, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:00:52,072 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:00:52,075 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:46564, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:00:52,076 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x3257e4ca to 127.0.0.1:59048 2025-01-08T05:00:52,076 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:00:52,077 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x6046f146 to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@7d7957e7 2025-01-08T05:00:52,082 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@bffb5be, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:00:52,083 DEBUG [hconnection-0x3e79c001-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:00:52,084 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:46574, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:00:52,087 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x6046f146 to 127.0.0.1:59048 2025-01-08T05:00:52,087 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:00:52,087 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemStateWithMergeRegion], kv [jenkins: RWXCA] 2025-01-08T05:00:52,088 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T05:00:52,089 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=104, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } 2025-01-08T05:00:52,089 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 }, snapshot procedure id = 104 2025-01-08T05:00:52,090 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T05:00:52,091 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=104 2025-01-08T05:00:52,091 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T05:00:52,093 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T05:00:52,100 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742127_1303 (size=215) 2025-01-08T05:00:52,100 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742127_1303 (size=215) 2025-01-08T05:00:52,100 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742127_1303 (size=215) 2025-01-08T05:00:52,102 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T05:00:52,102 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=105, ppid=104, state=RUNNABLE; SnapshotRegionProcedure 762016803bad87dc082747011636b848}, {pid=106, ppid=104, state=RUNNABLE; SnapshotRegionProcedure 8ceb1f1fe1c86dfb459738a7200e1333}] 2025-01-08T05:00:52,103 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=106, ppid=104, state=RUNNABLE; SnapshotRegionProcedure 8ceb1f1fe1c86dfb459738a7200e1333 2025-01-08T05:00:52,103 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=105, ppid=104, state=RUNNABLE; SnapshotRegionProcedure 762016803bad87dc082747011636b848 2025-01-08T05:00:52,191 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=104 2025-01-08T05:00:52,254 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:00:52,254 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T05:00:52,255 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=37967 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=105 2025-01-08T05:00:52,255 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37583 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=106 2025-01-08T05:00:52,255 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=105}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848. 2025-01-08T05:00:52,255 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=106}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333. 2025-01-08T05:00:52,255 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=106}] regionserver.HRegion(2538): Flush status journal for 8ceb1f1fe1c86dfb459738a7200e1333: 2025-01-08T05:00:52,255 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=105}] regionserver.HRegion(2538): Flush status journal for 762016803bad87dc082747011636b848: 2025-01-08T05:00:52,255 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=106}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333. for emptySnaptb0-testExportFileSystemStateWithMergeRegion completed. 2025-01-08T05:00:52,255 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=105}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848. for emptySnaptb0-testExportFileSystemStateWithMergeRegion completed. 2025-01-08T05:00:52,255 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=106}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333.' region-info for snapshot=emptySnaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T05:00:52,255 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=106}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T05:00:52,255 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=105}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848.' region-info for snapshot=emptySnaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T05:00:52,255 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=106}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T05:00:52,255 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=105}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T05:00:52,255 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=105}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T05:00:52,263 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742128_1304 (size=86) 2025-01-08T05:00:52,264 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742128_1304 (size=86) 2025-01-08T05:00:52,264 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742129_1305 (size=86) 2025-01-08T05:00:52,265 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742128_1304 (size=86) 2025-01-08T05:00:52,265 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=105}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848. 2025-01-08T05:00:52,265 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=105}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=105 2025-01-08T05:00:52,265 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742129_1305 (size=86) 2025-01-08T05:00:52,265 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742129_1305 (size=86) 2025-01-08T05:00:52,265 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=106}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333. 2025-01-08T05:00:52,265 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=106}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=106 2025-01-08T05:00:52,265 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=105 2025-01-08T05:00:52,265 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportFileSystemStateWithMergeRegion on region 762016803bad87dc082747011636b848 2025-01-08T05:00:52,265 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=106 2025-01-08T05:00:52,266 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportFileSystemStateWithMergeRegion on region 8ceb1f1fe1c86dfb459738a7200e1333 2025-01-08T05:00:52,266 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=105, ppid=104, state=RUNNABLE; SnapshotRegionProcedure 762016803bad87dc082747011636b848 2025-01-08T05:00:52,266 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=106, ppid=104, state=RUNNABLE; SnapshotRegionProcedure 8ceb1f1fe1c86dfb459738a7200e1333 2025-01-08T05:00:52,267 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=105, ppid=104, state=SUCCESS; SnapshotRegionProcedure 762016803bad87dc082747011636b848 in 164 msec 2025-01-08T05:00:52,268 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=106, resume processing ppid=104 2025-01-08T05:00:52,268 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=106, ppid=104, state=SUCCESS; SnapshotRegionProcedure 8ceb1f1fe1c86dfb459738a7200e1333 in 164 msec 2025-01-08T05:00:52,268 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T05:00:52,269 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T05:00:52,270 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T05:00:52,270 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T05:00:52,270 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:00:52,271 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(328): No files under family: cf 2025-01-08T05:00:52,279 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742130_1306 (size=78) 2025-01-08T05:00:52,280 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742130_1306 (size=78) 2025-01-08T05:00:52,281 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742130_1306 (size=78) 2025-01-08T05:00:52,282 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T05:00:52,282 DEBUG [PEWorker-2 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for emptySnaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T05:00:52,282 DEBUG [PEWorker-2 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/emptySnaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T05:00:52,291 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742131_1307 (size=713) 2025-01-08T05:00:52,291 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742131_1307 (size=713) 2025-01-08T05:00:52,291 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742131_1307 (size=713) 2025-01-08T05:00:52,296 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T05:00:52,301 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T05:00:52,301 DEBUG [PEWorker-2 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/emptySnaptb0-testExportFileSystemStateWithMergeRegion to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/emptySnaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T05:00:52,302 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T05:00:52,303 DEBUG [PEWorker-2 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 }, snapshot procedure id = 104 2025-01-08T05:00:52,304 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=104, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } in 215 msec 2025-01-08T05:00:52,392 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=104 2025-01-08T05:00:52,393 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion, procId: 104 completed 2025-01-08T05:00:52,395 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37967 {}] regionserver.HRegion(8254): writing data to region testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T05:00:52,396 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37583 {}] regionserver.HRegion(8254): writing data to region testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T05:00:52,399 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T05:00:52,399 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848. 2025-01-08T05:00:52,399 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T05:00:52,409 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } 2025-01-08T05:00:52,409 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736312452409 (current time:1736312452409). 2025-01-08T05:00:52,409 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T05:00:52,409 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testExportFileSystemStateWithMergeRegion VERSION not specified, setting to 2 2025-01-08T05:00:52,409 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T05:00:52,410 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x6e2c3d03 to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@2f39bf8c 2025-01-08T05:00:52,413 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@32e938de, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:00:52,415 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:00:52,416 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:46588, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:00:52,417 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x6e2c3d03 to 127.0.0.1:59048 2025-01-08T05:00:52,417 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:00:52,418 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x1d69ce93 to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@7f4ed76d 2025-01-08T05:00:52,421 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@49a2f92, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:00:52,426 DEBUG [hconnection-0x7d45745-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:00:52,427 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:46602, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:00:52,430 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x1d69ce93 to 127.0.0.1:59048 2025-01-08T05:00:52,430 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:00:52,430 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemStateWithMergeRegion], kv [jenkins: RWXCA] 2025-01-08T05:00:52,431 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T05:00:52,432 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=107, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } 2025-01-08T05:00:52,432 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 }, snapshot procedure id = 107 2025-01-08T05:00:52,433 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T05:00:52,433 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=107 2025-01-08T05:00:52,433 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T05:00:52,435 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T05:00:52,441 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742132_1308 (size=210) 2025-01-08T05:00:52,441 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742132_1308 (size=210) 2025-01-08T05:00:52,441 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742132_1308 (size=210) 2025-01-08T05:00:52,443 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T05:00:52,443 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=108, ppid=107, state=RUNNABLE; SnapshotRegionProcedure 762016803bad87dc082747011636b848}, {pid=109, ppid=107, state=RUNNABLE; SnapshotRegionProcedure 8ceb1f1fe1c86dfb459738a7200e1333}] 2025-01-08T05:00:52,443 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=109, ppid=107, state=RUNNABLE; SnapshotRegionProcedure 8ceb1f1fe1c86dfb459738a7200e1333 2025-01-08T05:00:52,444 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=108, ppid=107, state=RUNNABLE; SnapshotRegionProcedure 762016803bad87dc082747011636b848 2025-01-08T05:00:52,534 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=107 2025-01-08T05:00:52,594 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:00:52,595 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37967 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=108 2025-01-08T05:00:52,596 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848. 2025-01-08T05:00:52,596 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.HRegion(2837): Flushing 762016803bad87dc082747011636b848 1/1 column families, dataSize=132 B heapSize=544 B 2025-01-08T05:00:52,596 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T05:00:52,599 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37583 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=109 2025-01-08T05:00:52,599 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333. 2025-01-08T05:00:52,599 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.HRegion(2837): Flushing 8ceb1f1fe1c86dfb459738a7200e1333 1/1 column families, dataSize=3.13 KB heapSize=7 KB 2025-01-08T05:00:52,616 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e202501087e1ddbc52f9a4cba91dd7aa7111f6827_762016803bad87dc082747011636b848 is 71, key is 0538563748cb7a0501df6d7bd5c58df4/cf:q/1736312452395/Put/seqid=0 2025-01-08T05:00:52,622 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742133_1309 (size=5032) 2025-01-08T05:00:52,622 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742133_1309 (size=5032) 2025-01-08T05:00:52,624 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b2025010845a19cd429654dad94fb8e4d1c4f5cc4_8ceb1f1fe1c86dfb459738a7200e1333 is 71, key is 13371bd423027c4183861cc3942481c1/cf:q/1736312452396/Put/seqid=0 2025-01-08T05:00:52,625 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742133_1309 (size=5032) 2025-01-08T05:00:52,626 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:00:52,630 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742134_1310 (size=8241) 2025-01-08T05:00:52,630 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742134_1310 (size=8241) 2025-01-08T05:00:52,630 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742134_1310 (size=8241) 2025-01-08T05:00:52,631 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:00:52,631 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e202501087e1ddbc52f9a4cba91dd7aa7111f6827_762016803bad87dc082747011636b848 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/d41d8cd98f00b204e9800998ecf8427e202501087e1ddbc52f9a4cba91dd7aa7111f6827_762016803bad87dc082747011636b848 2025-01-08T05:00:52,632 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/762016803bad87dc082747011636b848/.tmp/cf/0b3ad9b3094546f082040c46a082079c, store: [table=testtb-testExportFileSystemStateWithMergeRegion family=cf region=762016803bad87dc082747011636b848] 2025-01-08T05:00:52,632 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/762016803bad87dc082747011636b848/.tmp/cf/0b3ad9b3094546f082040c46a082079c is 224, key is 0b9c8fdef15ca2e56948b1d513a670706/cf:q/1736312452395/Put/seqid=0 2025-01-08T05:00:52,635 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b2025010845a19cd429654dad94fb8e4d1c4f5cc4_8ceb1f1fe1c86dfb459738a7200e1333 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/c4ca4238a0b923820dcc509a6f75849b2025010845a19cd429654dad94fb8e4d1c4f5cc4_8ceb1f1fe1c86dfb459738a7200e1333 2025-01-08T05:00:52,636 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/8ceb1f1fe1c86dfb459738a7200e1333/.tmp/cf/ad6eb7d3730a4aa682e10bc5d132a8ee, store: [table=testtb-testExportFileSystemStateWithMergeRegion family=cf region=8ceb1f1fe1c86dfb459738a7200e1333] 2025-01-08T05:00:52,636 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/8ceb1f1fe1c86dfb459738a7200e1333/.tmp/cf/ad6eb7d3730a4aa682e10bc5d132a8ee is 224, key is 1babb27669f2fc34c0dbe39439c8d8f3f/cf:q/1736312452396/Put/seqid=0 2025-01-08T05:00:52,639 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742135_1311 (size=5754) 2025-01-08T05:00:52,640 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742135_1311 (size=5754) 2025-01-08T05:00:52,640 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742135_1311 (size=5754) 2025-01-08T05:00:52,640 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=132, hasBloomFilter=true, into tmp file hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/762016803bad87dc082747011636b848/.tmp/cf/0b3ad9b3094546f082040c46a082079c 2025-01-08T05:00:52,645 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/762016803bad87dc082747011636b848/.tmp/cf/0b3ad9b3094546f082040c46a082079c as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/762016803bad87dc082747011636b848/cf/0b3ad9b3094546f082040c46a082079c 2025-01-08T05:00:52,649 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/762016803bad87dc082747011636b848/cf/0b3ad9b3094546f082040c46a082079c, entries=2, sequenceid=6, filesize=5.6 K 2025-01-08T05:00:52,650 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.HRegion(3040): Finished flush of dataSize ~132 B/132, heapSize ~528 B/528, currentSize=0 B/0 for 762016803bad87dc082747011636b848 in 54ms, sequenceid=6, compaction requested=false 2025-01-08T05:00:52,650 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportFileSystemStateWithMergeRegion' 2025-01-08T05:00:52,650 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742136_1312 (size=15937) 2025-01-08T05:00:52,651 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742136_1312 (size=15937) 2025-01-08T05:00:52,651 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742136_1312 (size=15937) 2025-01-08T05:00:52,651 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.HRegion(2538): Flush status journal for 762016803bad87dc082747011636b848: 2025-01-08T05:00:52,651 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848. for snaptb0-testExportFileSystemStateWithMergeRegion completed. 2025-01-08T05:00:52,652 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=3.1 K, hasBloomFilter=true, into tmp file hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/8ceb1f1fe1c86dfb459738a7200e1333/.tmp/cf/ad6eb7d3730a4aa682e10bc5d132a8ee 2025-01-08T05:00:52,652 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848.' region-info for snapshot=snaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T05:00:52,652 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T05:00:52,652 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/762016803bad87dc082747011636b848/cf/0b3ad9b3094546f082040c46a082079c] hfiles 2025-01-08T05:00:52,652 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/762016803bad87dc082747011636b848/cf/0b3ad9b3094546f082040c46a082079c for snapshot=snaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T05:00:52,657 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/8ceb1f1fe1c86dfb459738a7200e1333/.tmp/cf/ad6eb7d3730a4aa682e10bc5d132a8ee as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/8ceb1f1fe1c86dfb459738a7200e1333/cf/ad6eb7d3730a4aa682e10bc5d132a8ee 2025-01-08T05:00:52,658 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742137_1313 (size=125) 2025-01-08T05:00:52,658 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742137_1313 (size=125) 2025-01-08T05:00:52,659 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742137_1313 (size=125) 2025-01-08T05:00:52,659 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848. 2025-01-08T05:00:52,659 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=108 2025-01-08T05:00:52,660 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=108 2025-01-08T05:00:52,660 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportFileSystemStateWithMergeRegion on region 762016803bad87dc082747011636b848 2025-01-08T05:00:52,660 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=108, ppid=107, state=RUNNABLE; SnapshotRegionProcedure 762016803bad87dc082747011636b848 2025-01-08T05:00:52,662 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=108, ppid=107, state=SUCCESS; SnapshotRegionProcedure 762016803bad87dc082747011636b848 in 218 msec 2025-01-08T05:00:52,663 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/8ceb1f1fe1c86dfb459738a7200e1333/cf/ad6eb7d3730a4aa682e10bc5d132a8ee, entries=48, sequenceid=6, filesize=15.6 K 2025-01-08T05:00:52,664 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.HRegion(3040): Finished flush of dataSize ~3.13 KB/3204, heapSize ~6.98 KB/7152, currentSize=0 B/0 for 8ceb1f1fe1c86dfb459738a7200e1333 in 64ms, sequenceid=6, compaction requested=false 2025-01-08T05:00:52,664 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.HRegion(2538): Flush status journal for 8ceb1f1fe1c86dfb459738a7200e1333: 2025-01-08T05:00:52,664 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333. for snaptb0-testExportFileSystemStateWithMergeRegion completed. 2025-01-08T05:00:52,664 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333.' region-info for snapshot=snaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T05:00:52,664 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T05:00:52,664 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/8ceb1f1fe1c86dfb459738a7200e1333/cf/ad6eb7d3730a4aa682e10bc5d132a8ee] hfiles 2025-01-08T05:00:52,664 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/8ceb1f1fe1c86dfb459738a7200e1333/cf/ad6eb7d3730a4aa682e10bc5d132a8ee for snapshot=snaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T05:00:52,669 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742138_1314 (size=125) 2025-01-08T05:00:52,670 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742138_1314 (size=125) 2025-01-08T05:00:52,670 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742138_1314 (size=125) 2025-01-08T05:00:52,670 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333. 2025-01-08T05:00:52,670 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=109 2025-01-08T05:00:52,671 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=109 2025-01-08T05:00:52,671 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportFileSystemStateWithMergeRegion on region 8ceb1f1fe1c86dfb459738a7200e1333 2025-01-08T05:00:52,671 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=109, ppid=107, state=RUNNABLE; SnapshotRegionProcedure 8ceb1f1fe1c86dfb459738a7200e1333 2025-01-08T05:00:52,673 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=109, resume processing ppid=107 2025-01-08T05:00:52,673 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=109, ppid=107, state=SUCCESS; SnapshotRegionProcedure 8ceb1f1fe1c86dfb459738a7200e1333 in 229 msec 2025-01-08T05:00:52,673 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T05:00:52,674 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T05:00:52,674 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T05:00:52,674 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T05:00:52,675 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:00:52,676 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/c4ca4238a0b923820dcc509a6f75849b2025010845a19cd429654dad94fb8e4d1c4f5cc4_8ceb1f1fe1c86dfb459738a7200e1333, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/d41d8cd98f00b204e9800998ecf8427e202501087e1ddbc52f9a4cba91dd7aa7111f6827_762016803bad87dc082747011636b848] hfiles 2025-01-08T05:00:52,676 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/c4ca4238a0b923820dcc509a6f75849b2025010845a19cd429654dad94fb8e4d1c4f5cc4_8ceb1f1fe1c86dfb459738a7200e1333 2025-01-08T05:00:52,676 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/d41d8cd98f00b204e9800998ecf8427e202501087e1ddbc52f9a4cba91dd7aa7111f6827_762016803bad87dc082747011636b848 2025-01-08T05:00:52,681 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742139_1315 (size=309) 2025-01-08T05:00:52,681 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742139_1315 (size=309) 2025-01-08T05:00:52,682 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742139_1315 (size=309) 2025-01-08T05:00:52,683 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T05:00:52,683 DEBUG [PEWorker-4 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T05:00:52,684 DEBUG [PEWorker-4 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T05:00:52,699 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742140_1316 (size=1023) 2025-01-08T05:00:52,700 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742140_1316 (size=1023) 2025-01-08T05:00:52,700 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742140_1316 (size=1023) 2025-01-08T05:00:52,704 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T05:00:52,710 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T05:00:52,710 DEBUG [PEWorker-4 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemStateWithMergeRegion to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T05:00:52,712 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T05:00:52,712 DEBUG [PEWorker-4 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 }, snapshot procedure id = 107 2025-01-08T05:00:52,713 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=107, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } in 281 msec 2025-01-08T05:00:52,735 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=107 2025-01-08T05:00:52,735 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion, procId: 107 completed 2025-01-08T05:00:52,754 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2025-01-08T05:00:52,756 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:46618, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2025-01-08T05:00:52,757 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37583 {}] regionserver.CompactSplit(323): Interrupting running compactions because user switched off compactions 2025-01-08T05:00:52,757 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2025-01-08T05:00:52,759 INFO [RS-EventLoopGroup-5-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:39138, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2025-01-08T05:00:52,759 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37967 {}] regionserver.CompactSplit(323): Interrupting running compactions because user switched off compactions 2025-01-08T05:00:52,759 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2025-01-08T05:00:52,761 INFO [RS-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:41048, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2025-01-08T05:00:52,761 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39801 {}] regionserver.CompactSplit(323): Interrupting running compactions because user switched off compactions 2025-01-08T05:00:52,763 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.3 create 'testtb-testExportFileSystemStateWithMergeRegion-1', {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T05:00:52,764 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=110, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:00:52,764 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=110, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T05:00:52,765 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:00:52,765 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "testtb-testExportFileSystemStateWithMergeRegion-1" procId is: 110 2025-01-08T05:00:52,765 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=110, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T05:00:52,765 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=110 2025-01-08T05:00:52,776 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742141_1317 (size=399) 2025-01-08T05:00:52,776 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742141_1317 (size=399) 2025-01-08T05:00:52,776 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742141_1317 (size=399) 2025-01-08T05:00:52,779 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 4e1ff0e98851e58decfd4b7963bd4305, NAME => 'testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452762.4e1ff0e98851e58decfd4b7963bd4305.', STARTKEY => '', ENDKEY => '2'}, tableDescriptor='testtb-testExportFileSystemStateWithMergeRegion-1', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:00:52,779 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => 747352f4d6112e9ff87f40579cee217d, NAME => 'testtb-testExportFileSystemStateWithMergeRegion-1,2,1736312452762.747352f4d6112e9ff87f40579cee217d.', STARTKEY => '2', ENDKEY => ''}, tableDescriptor='testtb-testExportFileSystemStateWithMergeRegion-1', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:00:52,786 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742143_1319 (size=85) 2025-01-08T05:00:52,786 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742143_1319 (size=85) 2025-01-08T05:00:52,787 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742143_1319 (size=85) 2025-01-08T05:00:52,787 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742142_1318 (size=85) 2025-01-08T05:00:52,787 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452762.4e1ff0e98851e58decfd4b7963bd4305.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:00:52,787 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-0 {}] regionserver.HRegion(1681): Closing 4e1ff0e98851e58decfd4b7963bd4305, disabling compactions & flushes 2025-01-08T05:00:52,787 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452762.4e1ff0e98851e58decfd4b7963bd4305. 2025-01-08T05:00:52,788 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452762.4e1ff0e98851e58decfd4b7963bd4305. 2025-01-08T05:00:52,788 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452762.4e1ff0e98851e58decfd4b7963bd4305. after waiting 0 ms 2025-01-08T05:00:52,788 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452762.4e1ff0e98851e58decfd4b7963bd4305. 2025-01-08T05:00:52,788 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452762.4e1ff0e98851e58decfd4b7963bd4305. 2025-01-08T05:00:52,788 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-0 {}] regionserver.HRegion(1635): Region close journal for 4e1ff0e98851e58decfd4b7963bd4305: 2025-01-08T05:00:52,788 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742142_1318 (size=85) 2025-01-08T05:00:52,788 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742142_1318 (size=85) 2025-01-08T05:00:52,789 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithMergeRegion-1,2,1736312452762.747352f4d6112e9ff87f40579cee217d.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:00:52,789 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-1 {}] regionserver.HRegion(1681): Closing 747352f4d6112e9ff87f40579cee217d, disabling compactions & flushes 2025-01-08T05:00:52,789 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithMergeRegion-1,2,1736312452762.747352f4d6112e9ff87f40579cee217d. 2025-01-08T05:00:52,789 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithMergeRegion-1,2,1736312452762.747352f4d6112e9ff87f40579cee217d. 2025-01-08T05:00:52,789 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithMergeRegion-1,2,1736312452762.747352f4d6112e9ff87f40579cee217d. after waiting 0 ms 2025-01-08T05:00:52,789 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithMergeRegion-1,2,1736312452762.747352f4d6112e9ff87f40579cee217d. 2025-01-08T05:00:52,789 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithMergeRegion-1,2,1736312452762.747352f4d6112e9ff87f40579cee217d. 2025-01-08T05:00:52,789 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-1 {}] regionserver.HRegion(1635): Region close journal for 747352f4d6112e9ff87f40579cee217d: 2025-01-08T05:00:52,790 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=110, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T05:00:52,790 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452762.4e1ff0e98851e58decfd4b7963bd4305.","families":{"info":[{"qualifier":"regioninfo","vlen":84,"tag":[],"timestamp":"1736312452790"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736312452790"}]},"ts":"1736312452790"} 2025-01-08T05:00:52,790 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportFileSystemStateWithMergeRegion-1,2,1736312452762.747352f4d6112e9ff87f40579cee217d.","families":{"info":[{"qualifier":"regioninfo","vlen":84,"tag":[],"timestamp":"1736312452790"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736312452790"}]},"ts":"1736312452790"} 2025-01-08T05:00:52,792 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T05:00:52,793 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=110, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T05:00:52,793 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion-1","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312452793"}]},"ts":"1736312452793"} 2025-01-08T05:00:52,794 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithMergeRegion-1, state=ENABLING in hbase:meta 2025-01-08T05:00:52,799 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(202): Hosts are {f28094b72178=0} racks are {/default-rack=0} 2025-01-08T05:00:52,800 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T05:00:52,800 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T05:00:52,800 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T05:00:52,800 INFO [PEWorker-3 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T05:00:52,800 INFO [PEWorker-3 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T05:00:52,800 INFO [PEWorker-3 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T05:00:52,800 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T05:00:52,800 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=111, ppid=110, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=4e1ff0e98851e58decfd4b7963bd4305, ASSIGN}, {pid=112, ppid=110, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=747352f4d6112e9ff87f40579cee217d, ASSIGN}] 2025-01-08T05:00:52,801 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=112, ppid=110, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=747352f4d6112e9ff87f40579cee217d, ASSIGN 2025-01-08T05:00:52,801 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=111, ppid=110, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=4e1ff0e98851e58decfd4b7963bd4305, ASSIGN 2025-01-08T05:00:52,802 INFO [PEWorker-5 {}] assignment.TransitRegionStateProcedure(264): Starting pid=111, ppid=110, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=4e1ff0e98851e58decfd4b7963bd4305, ASSIGN; state=OFFLINE, location=f28094b72178,37967,1736312288941; forceNewPlan=false, retain=false 2025-01-08T05:00:52,802 INFO [PEWorker-1 {}] assignment.TransitRegionStateProcedure(264): Starting pid=112, ppid=110, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=747352f4d6112e9ff87f40579cee217d, ASSIGN; state=OFFLINE, location=f28094b72178,37583,1736312288873; forceNewPlan=false, retain=false 2025-01-08T05:00:52,866 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=110 2025-01-08T05:00:52,952 INFO [f28094b72178:42195 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T05:00:52,952 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=111 updating hbase:meta row=4e1ff0e98851e58decfd4b7963bd4305, regionState=OPENING, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T05:00:52,952 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=112 updating hbase:meta row=747352f4d6112e9ff87f40579cee217d, regionState=OPENING, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T05:00:52,954 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=113, ppid=111, state=RUNNABLE; OpenRegionProcedure 4e1ff0e98851e58decfd4b7963bd4305, server=f28094b72178,37967,1736312288941}] 2025-01-08T05:00:52,955 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=114, ppid=112, state=RUNNABLE; OpenRegionProcedure 747352f4d6112e9ff87f40579cee217d, server=f28094b72178,37583,1736312288873}] 2025-01-08T05:00:53,067 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=110 2025-01-08T05:00:53,106 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T05:00:53,106 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:00:53,109 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] handler.AssignRegionHandler(135): Open testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452762.4e1ff0e98851e58decfd4b7963bd4305. 2025-01-08T05:00:53,109 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] handler.AssignRegionHandler(135): Open testtb-testExportFileSystemStateWithMergeRegion-1,2,1736312452762.747352f4d6112e9ff87f40579cee217d. 2025-01-08T05:00:53,109 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(7285): Opening region: {ENCODED => 4e1ff0e98851e58decfd4b7963bd4305, NAME => 'testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452762.4e1ff0e98851e58decfd4b7963bd4305.', STARTKEY => '', ENDKEY => '2'} 2025-01-08T05:00:53,109 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(7285): Opening region: {ENCODED => 747352f4d6112e9ff87f40579cee217d, NAME => 'testtb-testExportFileSystemStateWithMergeRegion-1,2,1736312452762.747352f4d6112e9ff87f40579cee217d.', STARTKEY => '2', ENDKEY => ''} 2025-01-08T05:00:53,109 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportFileSystemStateWithMergeRegion-1,2,1736312452762.747352f4d6112e9ff87f40579cee217d. service=AccessControlService 2025-01-08T05:00:53,109 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452762.4e1ff0e98851e58decfd4b7963bd4305. service=AccessControlService 2025-01-08T05:00:53,110 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T05:00:53,110 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T05:00:53,110 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportFileSystemStateWithMergeRegion-1 747352f4d6112e9ff87f40579cee217d 2025-01-08T05:00:53,110 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportFileSystemStateWithMergeRegion-1 4e1ff0e98851e58decfd4b7963bd4305 2025-01-08T05:00:53,110 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithMergeRegion-1,2,1736312452762.747352f4d6112e9ff87f40579cee217d.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:00:53,110 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452762.4e1ff0e98851e58decfd4b7963bd4305.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:00:53,110 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(7327): checking encryption for 747352f4d6112e9ff87f40579cee217d 2025-01-08T05:00:53,110 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(7327): checking encryption for 4e1ff0e98851e58decfd4b7963bd4305 2025-01-08T05:00:53,110 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(7330): checking classloading for 747352f4d6112e9ff87f40579cee217d 2025-01-08T05:00:53,110 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(7330): checking classloading for 4e1ff0e98851e58decfd4b7963bd4305 2025-01-08T05:00:53,111 INFO [StoreOpener-747352f4d6112e9ff87f40579cee217d-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 747352f4d6112e9ff87f40579cee217d 2025-01-08T05:00:53,111 INFO [StoreOpener-4e1ff0e98851e58decfd4b7963bd4305-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 4e1ff0e98851e58decfd4b7963bd4305 2025-01-08T05:00:53,113 INFO [StoreOpener-4e1ff0e98851e58decfd4b7963bd4305-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 4e1ff0e98851e58decfd4b7963bd4305 columnFamilyName cf 2025-01-08T05:00:53,113 INFO [StoreOpener-747352f4d6112e9ff87f40579cee217d-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 747352f4d6112e9ff87f40579cee217d columnFamilyName cf 2025-01-08T05:00:53,113 DEBUG [StoreOpener-4e1ff0e98851e58decfd4b7963bd4305-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:00:53,113 DEBUG [StoreOpener-747352f4d6112e9ff87f40579cee217d-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:00:53,113 INFO [StoreOpener-4e1ff0e98851e58decfd4b7963bd4305-1 {}] regionserver.HStore(327): Store=4e1ff0e98851e58decfd4b7963bd4305/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T05:00:53,113 INFO [StoreOpener-747352f4d6112e9ff87f40579cee217d-1 {}] regionserver.HStore(327): Store=747352f4d6112e9ff87f40579cee217d/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T05:00:53,114 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/747352f4d6112e9ff87f40579cee217d 2025-01-08T05:00:53,114 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/4e1ff0e98851e58decfd4b7963bd4305 2025-01-08T05:00:53,114 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/747352f4d6112e9ff87f40579cee217d 2025-01-08T05:00:53,114 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/4e1ff0e98851e58decfd4b7963bd4305 2025-01-08T05:00:53,116 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(1085): writing seq id for 747352f4d6112e9ff87f40579cee217d 2025-01-08T05:00:53,116 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(1085): writing seq id for 4e1ff0e98851e58decfd4b7963bd4305 2025-01-08T05:00:53,118 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/4e1ff0e98851e58decfd4b7963bd4305/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T05:00:53,118 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/747352f4d6112e9ff87f40579cee217d/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T05:00:53,118 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(1102): Opened 747352f4d6112e9ff87f40579cee217d; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=73431557, jitterRate=0.09421546757221222}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T05:00:53,118 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(1102): Opened 4e1ff0e98851e58decfd4b7963bd4305; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=63424528, jitterRate=-0.0549008846282959}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T05:00:53,119 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(1001): Region open journal for 747352f4d6112e9ff87f40579cee217d: 2025-01-08T05:00:53,119 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(1001): Region open journal for 4e1ff0e98851e58decfd4b7963bd4305: 2025-01-08T05:00:53,120 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportFileSystemStateWithMergeRegion-1,2,1736312452762.747352f4d6112e9ff87f40579cee217d., pid=114, masterSystemTime=1736312453106 2025-01-08T05:00:53,120 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452762.4e1ff0e98851e58decfd4b7963bd4305., pid=113, masterSystemTime=1736312453106 2025-01-08T05:00:53,121 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452762.4e1ff0e98851e58decfd4b7963bd4305. 2025-01-08T05:00:53,121 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] handler.AssignRegionHandler(164): Opened testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452762.4e1ff0e98851e58decfd4b7963bd4305. 2025-01-08T05:00:53,122 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=111 updating hbase:meta row=4e1ff0e98851e58decfd4b7963bd4305, regionState=OPEN, openSeqNum=2, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T05:00:53,122 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportFileSystemStateWithMergeRegion-1,2,1736312452762.747352f4d6112e9ff87f40579cee217d. 2025-01-08T05:00:53,122 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] handler.AssignRegionHandler(164): Opened testtb-testExportFileSystemStateWithMergeRegion-1,2,1736312452762.747352f4d6112e9ff87f40579cee217d. 2025-01-08T05:00:53,122 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=112 updating hbase:meta row=747352f4d6112e9ff87f40579cee217d, regionState=OPEN, openSeqNum=2, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T05:00:53,125 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=113, resume processing ppid=111 2025-01-08T05:00:53,125 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=113, ppid=111, state=SUCCESS; OpenRegionProcedure 4e1ff0e98851e58decfd4b7963bd4305, server=f28094b72178,37967,1736312288941 in 169 msec 2025-01-08T05:00:53,125 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=114, resume processing ppid=112 2025-01-08T05:00:53,125 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=114, ppid=112, state=SUCCESS; OpenRegionProcedure 747352f4d6112e9ff87f40579cee217d, server=f28094b72178,37583,1736312288873 in 169 msec 2025-01-08T05:00:53,126 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=111, ppid=110, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=4e1ff0e98851e58decfd4b7963bd4305, ASSIGN in 325 msec 2025-01-08T05:00:53,127 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=112, resume processing ppid=110 2025-01-08T05:00:53,127 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=112, ppid=110, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=747352f4d6112e9ff87f40579cee217d, ASSIGN in 325 msec 2025-01-08T05:00:53,127 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=110, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T05:00:53,127 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion-1","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312453127"}]},"ts":"1736312453127"} 2025-01-08T05:00:53,128 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithMergeRegion-1, state=ENABLED in hbase:meta 2025-01-08T05:00:53,130 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=110, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T05:00:53,131 DEBUG [PEWorker-1 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testExportFileSystemStateWithMergeRegion-1 jenkins: RWXCA 2025-01-08T05:00:53,132 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37583 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemStateWithMergeRegion-1], kv [jenkins: RWXCA] 2025-01-08T05:00:53,134 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:00:53,134 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:00:53,134 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:00:53,134 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:00:53,137 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T05:00:53,137 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion-1 with data PBUF\x0AW\x0A\x07jenkins\x12L\x08\x03"H\x0A<\x0A\x07default\x121testtb-testExportFileSystemStateWithMergeRegion-1 \x00 \x01 \x02 \x03 \x04 2025-01-08T05:00:53,137 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T05:00:53,137 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T05:00:53,137 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion-1 with data PBUF\x0AW\x0A\x07jenkins\x12L\x08\x03"H\x0A<\x0A\x07default\x121testtb-testExportFileSystemStateWithMergeRegion-1 \x00 \x01 \x02 \x03 \x04 2025-01-08T05:00:53,137 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion-1 with data PBUF\x0AW\x0A\x07jenkins\x12L\x08\x03"H\x0A<\x0A\x07default\x121testtb-testExportFileSystemStateWithMergeRegion-1 \x00 \x01 \x02 \x03 \x04 2025-01-08T05:00:53,138 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T05:00:53,138 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion-1 with data PBUF\x0AW\x0A\x07jenkins\x12L\x08\x03"H\x0A<\x0A\x07default\x121testtb-testExportFileSystemStateWithMergeRegion-1 \x00 \x01 \x02 \x03 \x04 2025-01-08T05:00:53,138 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=110, state=SUCCESS; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 in 374 msec 2025-01-08T05:00:53,368 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=110 2025-01-08T05:00:53,369 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion-1, procId: 110 completed 2025-01-08T05:00:53,387 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster$2(2234): Client=jenkins//172.17.0.3 merge regions [4e1ff0e98851e58decfd4b7963bd4305, 747352f4d6112e9ff87f40579cee217d] 2025-01-08T05:00:53,393 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=115, state=RUNNABLE:MERGE_TABLE_REGIONS_PREPARE; MergeTableRegionsProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, regions=[4e1ff0e98851e58decfd4b7963bd4305, 747352f4d6112e9ff87f40579cee217d], force=true 2025-01-08T05:00:53,394 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=115, state=RUNNABLE:MERGE_TABLE_REGIONS_PREPARE; MergeTableRegionsProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, regions=[4e1ff0e98851e58decfd4b7963bd4305, 747352f4d6112e9ff87f40579cee217d], force=true 2025-01-08T05:00:53,394 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=115, state=RUNNABLE:MERGE_TABLE_REGIONS_PREPARE; MergeTableRegionsProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, regions=[4e1ff0e98851e58decfd4b7963bd4305, 747352f4d6112e9ff87f40579cee217d], force=true 2025-01-08T05:00:53,394 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=115, state=RUNNABLE:MERGE_TABLE_REGIONS_PREPARE; MergeTableRegionsProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, regions=[4e1ff0e98851e58decfd4b7963bd4305, 747352f4d6112e9ff87f40579cee217d], force=true 2025-01-08T05:00:53,395 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=115 2025-01-08T05:00:53,407 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=116, ppid=115, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=4e1ff0e98851e58decfd4b7963bd4305, UNASSIGN}, {pid=117, ppid=115, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=747352f4d6112e9ff87f40579cee217d, UNASSIGN}] 2025-01-08T05:00:53,408 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=116, ppid=115, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=4e1ff0e98851e58decfd4b7963bd4305, UNASSIGN 2025-01-08T05:00:53,408 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=117, ppid=115, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=747352f4d6112e9ff87f40579cee217d, UNASSIGN 2025-01-08T05:00:53,408 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=116 updating hbase:meta row=4e1ff0e98851e58decfd4b7963bd4305, regionState=CLOSING, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T05:00:53,408 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=117 updating hbase:meta row=747352f4d6112e9ff87f40579cee217d, regionState=CLOSING, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T05:00:53,410 DEBUG [PEWorker-5 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: true: evictOnSplit: true: evictOnClose: false 2025-01-08T05:00:53,410 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=118, ppid=116, state=RUNNABLE; CloseRegionProcedure 4e1ff0e98851e58decfd4b7963bd4305, server=f28094b72178,37967,1736312288941}] 2025-01-08T05:00:53,410 DEBUG [PEWorker-2 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: true: evictOnSplit: true: evictOnClose: false 2025-01-08T05:00:53,411 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=119, ppid=117, state=RUNNABLE; CloseRegionProcedure 747352f4d6112e9ff87f40579cee217d, server=f28094b72178,37583,1736312288873}] 2025-01-08T05:00:53,498 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=115 2025-01-08T05:00:53,561 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:00:53,562 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] handler.UnassignRegionHandler(124): Close 4e1ff0e98851e58decfd4b7963bd4305 2025-01-08T05:00:53,562 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] handler.UnassignRegionHandler(138): Unassign region: split region: true: evictCache: true 2025-01-08T05:00:53,562 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegion(1681): Closing 4e1ff0e98851e58decfd4b7963bd4305, disabling compactions & flushes 2025-01-08T05:00:53,562 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452762.4e1ff0e98851e58decfd4b7963bd4305. 2025-01-08T05:00:53,562 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452762.4e1ff0e98851e58decfd4b7963bd4305. 2025-01-08T05:00:53,562 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T05:00:53,562 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452762.4e1ff0e98851e58decfd4b7963bd4305. after waiting 0 ms 2025-01-08T05:00:53,562 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452762.4e1ff0e98851e58decfd4b7963bd4305. 2025-01-08T05:00:53,562 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegion(2837): Flushing 4e1ff0e98851e58decfd4b7963bd4305 1/1 column families, dataSize=24 B heapSize=352 B 2025-01-08T05:00:53,563 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] handler.UnassignRegionHandler(124): Close 747352f4d6112e9ff87f40579cee217d 2025-01-08T05:00:53,563 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] handler.UnassignRegionHandler(138): Unassign region: split region: true: evictCache: true 2025-01-08T05:00:53,563 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegion(1681): Closing 747352f4d6112e9ff87f40579cee217d, disabling compactions & flushes 2025-01-08T05:00:53,563 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithMergeRegion-1,2,1736312452762.747352f4d6112e9ff87f40579cee217d. 2025-01-08T05:00:53,563 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithMergeRegion-1,2,1736312452762.747352f4d6112e9ff87f40579cee217d. 2025-01-08T05:00:53,563 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithMergeRegion-1,2,1736312452762.747352f4d6112e9ff87f40579cee217d. after waiting 0 ms 2025-01-08T05:00:53,563 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithMergeRegion-1,2,1736312452762.747352f4d6112e9ff87f40579cee217d. 2025-01-08T05:00:53,563 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegion(2837): Flushing 747352f4d6112e9ff87f40579cee217d 1/1 column families, dataSize=24 B heapSize=352 B 2025-01-08T05:00:53,580 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/4e1ff0e98851e58decfd4b7963bd4305/.tmp/cf/e7b558af0b6c41c7a0cc026545a9e6bb is 28, key is 1/cf:/1736312453372/Put/seqid=0 2025-01-08T05:00:53,584 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/747352f4d6112e9ff87f40579cee217d/.tmp/cf/594ade9bfa6147828e69c650755e989f is 28, key is 2/cf:/1736312453375/Put/seqid=0 2025-01-08T05:00:53,586 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742144_1320 (size=4945) 2025-01-08T05:00:53,586 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742144_1320 (size=4945) 2025-01-08T05:00:53,586 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742144_1320 (size=4945) 2025-01-08T05:00:53,587 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=24 B at sequenceid=5 (bloomFilter=true), to=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/4e1ff0e98851e58decfd4b7963bd4305/.tmp/cf/e7b558af0b6c41c7a0cc026545a9e6bb 2025-01-08T05:00:53,590 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742145_1321 (size=4945) 2025-01-08T05:00:53,590 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742145_1321 (size=4945) 2025-01-08T05:00:53,591 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742145_1321 (size=4945) 2025-01-08T05:00:53,591 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=24 B at sequenceid=5 (bloomFilter=true), to=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/747352f4d6112e9ff87f40579cee217d/.tmp/cf/594ade9bfa6147828e69c650755e989f 2025-01-08T05:00:53,593 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/4e1ff0e98851e58decfd4b7963bd4305/.tmp/cf/e7b558af0b6c41c7a0cc026545a9e6bb as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/4e1ff0e98851e58decfd4b7963bd4305/cf/e7b558af0b6c41c7a0cc026545a9e6bb 2025-01-08T05:00:53,596 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/747352f4d6112e9ff87f40579cee217d/.tmp/cf/594ade9bfa6147828e69c650755e989f as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/747352f4d6112e9ff87f40579cee217d/cf/594ade9bfa6147828e69c650755e989f 2025-01-08T05:00:53,598 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/4e1ff0e98851e58decfd4b7963bd4305/cf/e7b558af0b6c41c7a0cc026545a9e6bb, entries=1, sequenceid=5, filesize=4.8 K 2025-01-08T05:00:53,599 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegion(3040): Finished flush of dataSize ~24 B/24, heapSize ~336 B/336, currentSize=0 B/0 for 4e1ff0e98851e58decfd4b7963bd4305 in 37ms, sequenceid=5, compaction requested=false 2025-01-08T05:00:53,599 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportFileSystemStateWithMergeRegion-1' 2025-01-08T05:00:53,601 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/747352f4d6112e9ff87f40579cee217d/cf/594ade9bfa6147828e69c650755e989f, entries=1, sequenceid=5, filesize=4.8 K 2025-01-08T05:00:53,601 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegion(3040): Finished flush of dataSize ~24 B/24, heapSize ~336 B/336, currentSize=0 B/0 for 747352f4d6112e9ff87f40579cee217d in 38ms, sequenceid=5, compaction requested=false 2025-01-08T05:00:53,607 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/4e1ff0e98851e58decfd4b7963bd4305/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2025-01-08T05:00:53,607 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:00:53,607 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452762.4e1ff0e98851e58decfd4b7963bd4305. 2025-01-08T05:00:53,607 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegion(1635): Region close journal for 4e1ff0e98851e58decfd4b7963bd4305: 2025-01-08T05:00:53,609 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] handler.UnassignRegionHandler(170): Closed 4e1ff0e98851e58decfd4b7963bd4305 2025-01-08T05:00:53,609 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/747352f4d6112e9ff87f40579cee217d/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2025-01-08T05:00:53,609 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=116 updating hbase:meta row=4e1ff0e98851e58decfd4b7963bd4305, regionState=CLOSED 2025-01-08T05:00:53,610 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:00:53,610 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithMergeRegion-1,2,1736312452762.747352f4d6112e9ff87f40579cee217d. 2025-01-08T05:00:53,610 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegion(1635): Region close journal for 747352f4d6112e9ff87f40579cee217d: 2025-01-08T05:00:53,612 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=117 updating hbase:meta row=747352f4d6112e9ff87f40579cee217d, regionState=CLOSED 2025-01-08T05:00:53,613 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] handler.UnassignRegionHandler(170): Closed 747352f4d6112e9ff87f40579cee217d 2025-01-08T05:00:53,613 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=118, resume processing ppid=116 2025-01-08T05:00:53,613 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=118, ppid=116, state=SUCCESS; CloseRegionProcedure 4e1ff0e98851e58decfd4b7963bd4305, server=f28094b72178,37967,1736312288941 in 201 msec 2025-01-08T05:00:53,614 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=116, ppid=115, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=4e1ff0e98851e58decfd4b7963bd4305, UNASSIGN in 206 msec 2025-01-08T05:00:53,615 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=119, resume processing ppid=117 2025-01-08T05:00:53,615 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=119, ppid=117, state=SUCCESS; CloseRegionProcedure 747352f4d6112e9ff87f40579cee217d, server=f28094b72178,37583,1736312288873 in 203 msec 2025-01-08T05:00:53,616 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=117, resume processing ppid=115 2025-01-08T05:00:53,616 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=117, ppid=115, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=747352f4d6112e9ff87f40579cee217d, UNASSIGN in 208 msec 2025-01-08T05:00:53,630 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742146_1322 (size=84) 2025-01-08T05:00:53,630 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742146_1322 (size=84) 2025-01-08T05:00:53,631 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742146_1322 (size=84) 2025-01-08T05:00:53,633 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:00:53,642 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742147_1323 (size=20) 2025-01-08T05:00:53,642 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742147_1323 (size=20) 2025-01-08T05:00:53,642 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742147_1323 (size=20) 2025-01-08T05:00:53,644 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:00:53,649 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742148_1324 (size=21) 2025-01-08T05:00:53,650 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742148_1324 (size=21) 2025-01-08T05:00:53,650 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742148_1324 (size=21) 2025-01-08T05:00:53,655 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742149_1325 (size=84) 2025-01-08T05:00:53,655 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742149_1325 (size=84) 2025-01-08T05:00:53,656 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742149_1325 (size=84) 2025-01-08T05:00:53,657 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:00:53,667 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a739dece698030682185693f1769a5eb/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=-1 2025-01-08T05:00:53,669 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452762.4e1ff0e98851e58decfd4b7963bd4305.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"9223372036854775807"}]},"ts":"9223372036854775807"} 2025-01-08T05:00:53,669 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion-1,2,1736312452762.747352f4d6112e9ff87f40579cee217d.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"9223372036854775807"}]},"ts":"9223372036854775807"} 2025-01-08T05:00:53,670 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":7,"row":"testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452763.a739dece698030682185693f1769a5eb.","families":{"info":[{"qualifier":"regioninfo","vlen":83,"tag":[],"timestamp":"9223372036854775807"},{"qualifier":"merge0000","vlen":84,"tag":[],"timestamp":"9223372036854775807"},{"qualifier":"merge0001","vlen":84,"tag":[],"timestamp":"9223372036854775807"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"9223372036854775807"}]},"ts":"9223372036854775807"} 2025-01-08T05:00:53,700 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=115 2025-01-08T05:00:53,707 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=120, ppid=115, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=a739dece698030682185693f1769a5eb, ASSIGN}] 2025-01-08T05:00:53,708 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=120, ppid=115, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=a739dece698030682185693f1769a5eb, ASSIGN 2025-01-08T05:00:53,709 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=120, ppid=115, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=a739dece698030682185693f1769a5eb, ASSIGN; state=MERGED, location=f28094b72178,37967,1736312288941; forceNewPlan=false, retain=false 2025-01-08T05:00:53,859 INFO [f28094b72178:42195 {}] balancer.BaseLoadBalancer(546): Reassigned 1 regions. 1 retained the pre-restart assignment. 2025-01-08T05:00:53,859 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=120 updating hbase:meta row=a739dece698030682185693f1769a5eb, regionState=OPENING, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T05:00:53,861 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=121, ppid=120, state=RUNNABLE; OpenRegionProcedure a739dece698030682185693f1769a5eb, server=f28094b72178,37967,1736312288941}] 2025-01-08T05:00:54,005 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=115 2025-01-08T05:00:54,013 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:00:54,017 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] handler.AssignRegionHandler(135): Open testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452763.a739dece698030682185693f1769a5eb. 2025-01-08T05:00:54,017 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(7285): Opening region: {ENCODED => a739dece698030682185693f1769a5eb, NAME => 'testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452763.a739dece698030682185693f1769a5eb.', STARTKEY => '', ENDKEY => ''} 2025-01-08T05:00:54,017 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452763.a739dece698030682185693f1769a5eb. service=AccessControlService 2025-01-08T05:00:54,017 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T05:00:54,018 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportFileSystemStateWithMergeRegion-1 a739dece698030682185693f1769a5eb 2025-01-08T05:00:54,018 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452763.a739dece698030682185693f1769a5eb.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:00:54,018 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(7327): checking encryption for a739dece698030682185693f1769a5eb 2025-01-08T05:00:54,018 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(7330): checking classloading for a739dece698030682185693f1769a5eb 2025-01-08T05:00:54,019 INFO [StoreOpener-a739dece698030682185693f1769a5eb-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region a739dece698030682185693f1769a5eb 2025-01-08T05:00:54,020 INFO [StoreOpener-a739dece698030682185693f1769a5eb-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region a739dece698030682185693f1769a5eb columnFamilyName cf 2025-01-08T05:00:54,020 DEBUG [StoreOpener-a739dece698030682185693f1769a5eb-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:00:54,042 DEBUG [StoreOpener-a739dece698030682185693f1769a5eb-1 {}] regionserver.StoreEngine(277): loaded hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a739dece698030682185693f1769a5eb/cf/594ade9bfa6147828e69c650755e989f.747352f4d6112e9ff87f40579cee217d->hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/747352f4d6112e9ff87f40579cee217d/cf/594ade9bfa6147828e69c650755e989f-top 2025-01-08T05:00:54,049 DEBUG [StoreOpener-a739dece698030682185693f1769a5eb-1 {}] regionserver.StoreEngine(277): loaded hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a739dece698030682185693f1769a5eb/cf/e7b558af0b6c41c7a0cc026545a9e6bb.4e1ff0e98851e58decfd4b7963bd4305->hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/4e1ff0e98851e58decfd4b7963bd4305/cf/e7b558af0b6c41c7a0cc026545a9e6bb-top 2025-01-08T05:00:54,050 INFO [StoreOpener-a739dece698030682185693f1769a5eb-1 {}] regionserver.HStore(327): Store=a739dece698030682185693f1769a5eb/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T05:00:54,051 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a739dece698030682185693f1769a5eb 2025-01-08T05:00:54,052 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a739dece698030682185693f1769a5eb 2025-01-08T05:00:54,055 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(1085): writing seq id for a739dece698030682185693f1769a5eb 2025-01-08T05:00:54,055 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(1102): Opened a739dece698030682185693f1769a5eb; next sequenceid=9; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=59721616, jitterRate=-0.11007857322692871}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T05:00:54,056 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(1001): Region open journal for a739dece698030682185693f1769a5eb: 2025-01-08T05:00:54,057 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452763.a739dece698030682185693f1769a5eb., pid=121, masterSystemTime=1736312454013 2025-01-08T05:00:54,058 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.CompactSplit(342): Ignoring compaction request for testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452763.a739dece698030682185693f1769a5eb.,because compaction is disabled. 2025-01-08T05:00:54,059 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452763.a739dece698030682185693f1769a5eb. 2025-01-08T05:00:54,059 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] handler.AssignRegionHandler(164): Opened testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452763.a739dece698030682185693f1769a5eb. 2025-01-08T05:00:54,060 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=120 updating hbase:meta row=a739dece698030682185693f1769a5eb, regionState=OPEN, openSeqNum=9, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T05:00:54,063 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=121, resume processing ppid=120 2025-01-08T05:00:54,063 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=121, ppid=120, state=SUCCESS; OpenRegionProcedure a739dece698030682185693f1769a5eb, server=f28094b72178,37967,1736312288941 in 201 msec 2025-01-08T05:00:54,065 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=120, resume processing ppid=115 2025-01-08T05:00:54,065 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=120, ppid=115, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=a739dece698030682185693f1769a5eb, ASSIGN in 356 msec 2025-01-08T05:00:54,067 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=115, state=SUCCESS; MergeTableRegionsProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, regions=[4e1ff0e98851e58decfd4b7963bd4305, 747352f4d6112e9ff87f40579cee217d], force=true in 676 msec 2025-01-08T05:00:54,506 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=115 2025-01-08T05:00:54,506 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: MERGE_REGIONS, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion-1, procId: 115 completed 2025-01-08T05:00:54,507 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } 2025-01-08T05:00:54,507 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736312454507 (current time:1736312454507). 2025-01-08T05:00:54,507 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T05:00:54,507 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testExportFileSystemStateWithMergeRegion-1 VERSION not specified, setting to 2 2025-01-08T05:00:54,507 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T05:00:54,507 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x6daa618d to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@204086cf 2025-01-08T05:00:54,511 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2e9e8105, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:00:54,512 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:00:54,513 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:38674, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:00:54,514 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x6daa618d to 127.0.0.1:59048 2025-01-08T05:00:54,514 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:00:54,515 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x4a8710be to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@6ed591df 2025-01-08T05:00:54,518 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@3499a3a9, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:00:54,520 DEBUG [hconnection-0x3d02ce7a-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:00:54,520 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:38690, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:00:54,523 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x4a8710be to 127.0.0.1:59048 2025-01-08T05:00:54,523 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:00:54,523 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemStateWithMergeRegion-1], kv [jenkins: RWXCA] 2025-01-08T05:00:54,524 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T05:00:54,524 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=122, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } 2025-01-08T05:00:54,525 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 }, snapshot procedure id = 122 2025-01-08T05:00:54,525 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T05:00:54,525 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=122 2025-01-08T05:00:54,526 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T05:00:54,528 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T05:00:54,533 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742150_1326 (size=216) 2025-01-08T05:00:54,533 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742150_1326 (size=216) 2025-01-08T05:00:54,534 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742150_1326 (size=216) 2025-01-08T05:00:54,535 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T05:00:54,535 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=123, ppid=122, state=RUNNABLE; SnapshotRegionProcedure a739dece698030682185693f1769a5eb}] 2025-01-08T05:00:54,536 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=123, ppid=122, state=RUNNABLE; SnapshotRegionProcedure a739dece698030682185693f1769a5eb 2025-01-08T05:00:54,626 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=122 2025-01-08T05:00:54,687 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:00:54,688 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=37967 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=123 2025-01-08T05:00:54,688 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452763.a739dece698030682185693f1769a5eb. 2025-01-08T05:00:54,688 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] regionserver.HRegion(2538): Flush status journal for a739dece698030682185693f1769a5eb: 2025-01-08T05:00:54,688 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452763.a739dece698030682185693f1769a5eb. for snaptb0-testExportFileSystemStateWithMergeRegion-1 completed. 2025-01-08T05:00:54,688 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452763.a739dece698030682185693f1769a5eb.' region-info for snapshot=snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:00:54,688 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T05:00:54,688 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a739dece698030682185693f1769a5eb/cf/594ade9bfa6147828e69c650755e989f.747352f4d6112e9ff87f40579cee217d->hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/747352f4d6112e9ff87f40579cee217d/cf/594ade9bfa6147828e69c650755e989f-top, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a739dece698030682185693f1769a5eb/cf/e7b558af0b6c41c7a0cc026545a9e6bb.4e1ff0e98851e58decfd4b7963bd4305->hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/4e1ff0e98851e58decfd4b7963bd4305/cf/e7b558af0b6c41c7a0cc026545a9e6bb-top] hfiles 2025-01-08T05:00:54,689 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] snapshot.SnapshotManifest(265): Adding reference for file (1/2): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a739dece698030682185693f1769a5eb/cf/594ade9bfa6147828e69c650755e989f.747352f4d6112e9ff87f40579cee217d for snapshot=snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:00:54,689 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] snapshot.SnapshotManifest(265): Adding reference for file (2/2): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a739dece698030682185693f1769a5eb/cf/e7b558af0b6c41c7a0cc026545a9e6bb.4e1ff0e98851e58decfd4b7963bd4305 for snapshot=snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:00:54,696 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742151_1327 (size=269) 2025-01-08T05:00:54,696 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742151_1327 (size=269) 2025-01-08T05:00:54,696 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742151_1327 (size=269) 2025-01-08T05:00:54,697 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452763.a739dece698030682185693f1769a5eb. 2025-01-08T05:00:54,697 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=123 2025-01-08T05:00:54,697 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=123 2025-01-08T05:00:54,697 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportFileSystemStateWithMergeRegion-1 on region a739dece698030682185693f1769a5eb 2025-01-08T05:00:54,697 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=123, ppid=122, state=RUNNABLE; SnapshotRegionProcedure a739dece698030682185693f1769a5eb 2025-01-08T05:00:54,700 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=123, resume processing ppid=122 2025-01-08T05:00:54,700 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=123, ppid=122, state=SUCCESS; SnapshotRegionProcedure a739dece698030682185693f1769a5eb in 163 msec 2025-01-08T05:00:54,700 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T05:00:54,700 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T05:00:54,702 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T05:00:54,702 DEBUG [PEWorker-4 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:00:54,702 DEBUG [PEWorker-4 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:00:54,711 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742152_1328 (size=670) 2025-01-08T05:00:54,711 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742152_1328 (size=670) 2025-01-08T05:00:54,712 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742152_1328 (size=670) 2025-01-08T05:00:54,714 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T05:00:54,719 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T05:00:54,720 DEBUG [PEWorker-4 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemStateWithMergeRegion-1 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:00:54,721 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T05:00:54,721 DEBUG [PEWorker-4 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 }, snapshot procedure id = 122 2025-01-08T05:00:54,722 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=122, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } in 197 msec 2025-01-08T05:00:54,827 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=122 2025-01-08T05:00:54,827 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion-1, procId: 122 completed 2025-01-08T05:00:54,828 INFO [Time-limited test {}] snapshot.TestExportSnapshot(468): HDFS export destination path: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312454827 2025-01-08T05:00:54,828 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=hdfs://localhost:45939, tgtDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312454827, rawTgtDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312454827, srcFsUri=hdfs://localhost:45939, srcDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:00:54,858 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:45939, inputRoot=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:00:54,858 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1149249371_22, ugi=jenkins (auth:SIMPLE)]], outputRoot=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312454827, skipTmp=false, initialOutputSnapshotDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312454827/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:00:54,860 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T05:00:54,865 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312454827/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:00:54,883 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742153_1329 (size=216) 2025-01-08T05:00:54,883 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742153_1329 (size=216) 2025-01-08T05:00:54,883 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742154_1330 (size=670) 2025-01-08T05:00:54,883 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742153_1329 (size=216) 2025-01-08T05:00:54,883 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742154_1330 (size=670) 2025-01-08T05:00:54,884 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742154_1330 (size=670) 2025-01-08T05:00:54,886 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:54,886 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:54,886 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:54,886 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:55,069 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0005_000001 (auth:SIMPLE) from 127.0.0.1:34436 2025-01-08T05:00:55,085 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_1/usercache/jenkins/appcache/application_1736312298603_0005/container_1736312298603_0005_01_000001/launch_container.sh] 2025-01-08T05:00:55,085 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_1/usercache/jenkins/appcache/application_1736312298603_0005/container_1736312298603_0005_01_000001/container_tokens] 2025-01-08T05:00:55,085 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_1/usercache/jenkins/appcache/application_1736312298603_0005/container_1736312298603_0005_01_000001/sysfs] 2025-01-08T05:00:56,009 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop-15595605909398982409.jar 2025-01-08T05:00:56,009 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:56,009 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:56,098 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop-9114897064428763168.jar 2025-01-08T05:00:56,098 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:56,099 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:56,099 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:56,099 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:56,100 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:56,100 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T05:00:56,100 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T05:00:56,100 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T05:00:56,101 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T05:00:56,101 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T05:00:56,101 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T05:00:56,102 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T05:00:56,102 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T05:00:56,102 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T05:00:56,103 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T05:00:56,103 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T05:00:56,103 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T05:00:56,103 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T05:00:56,104 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:00:56,104 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:00:56,104 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T05:00:56,105 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:00:56,105 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:00:56,105 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T05:00:56,106 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T05:00:56,216 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742155_1331 (size=127628) 2025-01-08T05:00:56,218 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742155_1331 (size=127628) 2025-01-08T05:00:56,218 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742155_1331 (size=127628) 2025-01-08T05:00:56,258 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742156_1332 (size=2172137) 2025-01-08T05:00:56,258 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742156_1332 (size=2172137) 2025-01-08T05:00:56,259 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742156_1332 (size=2172137) 2025-01-08T05:00:56,296 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742157_1333 (size=213228) 2025-01-08T05:00:56,296 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742157_1333 (size=213228) 2025-01-08T05:00:56,297 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742157_1333 (size=213228) 2025-01-08T05:00:56,321 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742158_1334 (size=1877034) 2025-01-08T05:00:56,321 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742158_1334 (size=1877034) 2025-01-08T05:00:56,322 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742158_1334 (size=1877034) 2025-01-08T05:00:56,347 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742159_1335 (size=533455) 2025-01-08T05:00:56,347 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742159_1335 (size=533455) 2025-01-08T05:00:56,348 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742159_1335 (size=533455) 2025-01-08T05:00:56,371 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742160_1336 (size=451756) 2025-01-08T05:00:56,371 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742160_1336 (size=451756) 2025-01-08T05:00:56,372 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742160_1336 (size=451756) 2025-01-08T05:00:56,420 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742161_1337 (size=7280644) 2025-01-08T05:00:56,420 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742161_1337 (size=7280644) 2025-01-08T05:00:56,421 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742161_1337 (size=7280644) 2025-01-08T05:00:56,447 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742162_1338 (size=4188619) 2025-01-08T05:00:56,447 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742162_1338 (size=4188619) 2025-01-08T05:00:56,447 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742162_1338 (size=4188619) 2025-01-08T05:00:56,472 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742163_1339 (size=20406) 2025-01-08T05:00:56,473 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742163_1339 (size=20406) 2025-01-08T05:00:56,474 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742163_1339 (size=20406) 2025-01-08T05:00:56,501 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742164_1340 (size=75495) 2025-01-08T05:00:56,501 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742164_1340 (size=75495) 2025-01-08T05:00:56,502 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742164_1340 (size=75495) 2025-01-08T05:00:56,528 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742165_1341 (size=45609) 2025-01-08T05:00:56,528 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742165_1341 (size=45609) 2025-01-08T05:00:56,528 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742165_1341 (size=45609) 2025-01-08T05:00:56,555 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742166_1342 (size=110084) 2025-01-08T05:00:56,556 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742166_1342 (size=110084) 2025-01-08T05:00:56,556 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742166_1342 (size=110084) 2025-01-08T05:00:56,590 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742167_1343 (size=1323991) 2025-01-08T05:00:56,590 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742167_1343 (size=1323991) 2025-01-08T05:00:56,591 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742167_1343 (size=1323991) 2025-01-08T05:00:56,598 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742168_1344 (size=23076) 2025-01-08T05:00:56,599 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742168_1344 (size=23076) 2025-01-08T05:00:56,599 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742168_1344 (size=23076) 2025-01-08T05:00:56,613 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742169_1345 (size=126803) 2025-01-08T05:00:56,613 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742169_1345 (size=126803) 2025-01-08T05:00:56,613 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742169_1345 (size=126803) 2025-01-08T05:00:56,628 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742170_1346 (size=322274) 2025-01-08T05:00:56,629 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742170_1346 (size=322274) 2025-01-08T05:00:56,629 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742170_1346 (size=322274) 2025-01-08T05:00:56,682 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T05:00:56,707 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742171_1347 (size=1832290) 2025-01-08T05:00:56,707 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742171_1347 (size=1832290) 2025-01-08T05:00:56,709 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742171_1347 (size=1832290) 2025-01-08T05:00:56,762 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742172_1348 (size=30081) 2025-01-08T05:00:56,762 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742172_1348 (size=30081) 2025-01-08T05:00:56,763 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742172_1348 (size=30081) 2025-01-08T05:00:56,804 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742173_1349 (size=53616) 2025-01-08T05:00:56,806 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742173_1349 (size=53616) 2025-01-08T05:00:56,806 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742173_1349 (size=53616) 2025-01-08T05:00:56,884 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742174_1350 (size=29229) 2025-01-08T05:00:56,885 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742174_1350 (size=29229) 2025-01-08T05:00:56,886 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742174_1350 (size=29229) 2025-01-08T05:00:56,920 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742175_1351 (size=169089) 2025-01-08T05:00:56,920 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742175_1351 (size=169089) 2025-01-08T05:00:56,921 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742175_1351 (size=169089) 2025-01-08T05:00:56,997 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742176_1352 (size=5175431) 2025-01-08T05:00:56,998 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742176_1352 (size=5175431) 2025-01-08T05:00:56,998 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742176_1352 (size=5175431) 2025-01-08T05:00:57,103 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742177_1353 (size=6350712) 2025-01-08T05:00:57,103 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742177_1353 (size=6350712) 2025-01-08T05:00:57,104 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742177_1353 (size=6350712) 2025-01-08T05:00:57,205 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742178_1354 (size=136454) 2025-01-08T05:00:57,208 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742178_1354 (size=136454) 2025-01-08T05:00:57,216 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742178_1354 (size=136454) 2025-01-08T05:00:57,309 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742179_1355 (size=907470) 2025-01-08T05:00:57,309 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742179_1355 (size=907470) 2025-01-08T05:00:57,309 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742179_1355 (size=907470) 2025-01-08T05:00:57,418 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742180_1356 (size=3317408) 2025-01-08T05:00:57,419 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742180_1356 (size=3317408) 2025-01-08T05:00:57,428 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742180_1356 (size=3317408) 2025-01-08T05:00:57,475 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742181_1357 (size=503880) 2025-01-08T05:00:57,477 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742181_1357 (size=503880) 2025-01-08T05:00:57,478 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742181_1357 (size=503880) 2025-01-08T05:00:57,594 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742182_1358 (size=4695811) 2025-01-08T05:00:57,596 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742182_1358 (size=4695811) 2025-01-08T05:00:57,596 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742182_1358 (size=4695811) 2025-01-08T05:00:57,598 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T05:00:57,600 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'snaptb0-testExportFileSystemStateWithMergeRegion-1' hfile list 2025-01-08T05:00:57,603 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(751): export split=0 size=9.7 K 2025-01-08T05:00:57,624 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742183_1359 (size=378) 2025-01-08T05:00:57,626 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742183_1359 (size=378) 2025-01-08T05:00:57,626 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742183_1359 (size=378) 2025-01-08T05:00:57,680 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742184_1360 (size=15) 2025-01-08T05:00:57,681 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742184_1360 (size=15) 2025-01-08T05:00:57,687 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742184_1360 (size=15) 2025-01-08T05:00:57,737 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742185_1361 (size=304938) 2025-01-08T05:00:57,738 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742185_1361 (size=304938) 2025-01-08T05:00:57,738 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742185_1361 (size=304938) 2025-01-08T05:00:57,766 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T05:00:57,766 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T05:00:58,092 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0006_000001 (auth:SIMPLE) from 127.0.0.1:34448 2025-01-08T05:00:58,345 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T05:00:58,345 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemStateWithMergeRegion Metrics about Tables on a single HBase RegionServer 2025-01-08T05:00:58,345 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:00:58,346 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemStateWithMergeRegion-1 Metrics about Tables on a single HBase RegionServer 2025-01-08T05:00:58,346 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testConsecutiveExports 2025-01-08T05:01:03,849 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T05:01:05,043 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0006_000001 (auth:SIMPLE) from 127.0.0.1:48430 2025-01-08T05:01:05,477 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742186_1362 (size=350612) 2025-01-08T05:01:05,477 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742186_1362 (size=350612) 2025-01-08T05:01:05,478 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742186_1362 (size=350612) 2025-01-08T05:01:05,791 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T05:01:07,584 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0006_000001 (auth:SIMPLE) from 127.0.0.1:38600 2025-01-08T05:01:12,403 DEBUG [master/f28094b72178:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 8ceb1f1fe1c86dfb459738a7200e1333 changed from -1.0 to 0.0, refreshing cache 2025-01-08T05:01:12,403 DEBUG [master/f28094b72178:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 762016803bad87dc082747011636b848 changed from -1.0 to 0.0, refreshing cache 2025-01-08T05:01:12,702 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742187_1363 (size=4945) 2025-01-08T05:01:12,703 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742187_1363 (size=4945) 2025-01-08T05:01:12,703 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742187_1363 (size=4945) 2025-01-08T05:01:12,776 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742188_1364 (size=4945) 2025-01-08T05:01:12,776 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742188_1364 (size=4945) 2025-01-08T05:01:12,776 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742188_1364 (size=4945) 2025-01-08T05:01:12,860 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742189_1365 (size=17474) 2025-01-08T05:01:12,861 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742189_1365 (size=17474) 2025-01-08T05:01:12,861 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742189_1365 (size=17474) 2025-01-08T05:01:12,878 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742190_1366 (size=482) 2025-01-08T05:01:12,878 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742190_1366 (size=482) 2025-01-08T05:01:12,879 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742190_1366 (size=482) 2025-01-08T05:01:12,928 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-1_2/usercache/jenkins/appcache/application_1736312298603_0006/container_1736312298603_0006_01_000002/launch_container.sh] 2025-01-08T05:01:12,928 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-1_2/usercache/jenkins/appcache/application_1736312298603_0006/container_1736312298603_0006_01_000002/container_tokens] 2025-01-08T05:01:12,928 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-1_2/usercache/jenkins/appcache/application_1736312298603_0006/container_1736312298603_0006_01_000002/sysfs] 2025-01-08T05:01:12,937 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742191_1367 (size=17474) 2025-01-08T05:01:12,938 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742191_1367 (size=17474) 2025-01-08T05:01:12,938 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742191_1367 (size=17474) 2025-01-08T05:01:12,978 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742192_1368 (size=350612) 2025-01-08T05:01:12,978 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742192_1368 (size=350612) 2025-01-08T05:01:12,979 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742192_1368 (size=350612) 2025-01-08T05:01:13,001 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0006_000001 (auth:SIMPLE) from 127.0.0.1:38602 2025-01-08T05:01:14,469 INFO [Time-limited test {}] snapshot.ExportSnapshot(1207): Finalize the Snapshot Export 2025-01-08T05:01:14,470 INFO [Time-limited test {}] snapshot.ExportSnapshot(1218): Verify the exported snapshot's expiration status and integrity. 2025-01-08T05:01:14,477 INFO [Time-limited test {}] snapshot.ExportSnapshot(1224): Export Completed: snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:01:14,478 INFO [Time-limited test {}] snapshot.TestExportSnapshot(362): Exported snapshot 2025-01-08T05:01:14,478 INFO [Time-limited test {}] snapshot.TestExportSnapshot(373): Verified filesystem state 2025-01-08T05:01:14,478 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1149249371_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1 at hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:01:14,479 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1/.snapshotinfo 2025-01-08T05:01:14,479 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1/data.manifest 2025-01-08T05:01:14,479 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1149249371_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312454827/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1 at hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312454827/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:01:14,479 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312454827/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1/.snapshotinfo 2025-01-08T05:01:14,479 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312454827/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1/data.manifest 2025-01-08T05:01:14,489 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:01:14,490 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.3 disable testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:01:14,491 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=124, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:01:14,493 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=124 2025-01-08T05:01:14,493 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion-1","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312474493"}]},"ts":"1736312474493"} 2025-01-08T05:01:14,494 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithMergeRegion-1, state=DISABLING in hbase:meta 2025-01-08T05:01:14,497 INFO [PEWorker-3 {}] procedure.DisableTableProcedure(284): Set testtb-testExportFileSystemStateWithMergeRegion-1 to state=DISABLING 2025-01-08T05:01:14,497 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=125, ppid=124, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1}] 2025-01-08T05:01:14,499 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=126, ppid=125, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=a739dece698030682185693f1769a5eb, UNASSIGN}] 2025-01-08T05:01:14,499 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=126, ppid=125, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=a739dece698030682185693f1769a5eb, UNASSIGN 2025-01-08T05:01:14,500 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=126 updating hbase:meta row=a739dece698030682185693f1769a5eb, regionState=CLOSING, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T05:01:14,501 DEBUG [PEWorker-5 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T05:01:14,501 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=127, ppid=126, state=RUNNABLE; CloseRegionProcedure a739dece698030682185693f1769a5eb, server=f28094b72178,37967,1736312288941}] 2025-01-08T05:01:14,593 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=124 2025-01-08T05:01:14,653 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:01:14,653 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] handler.UnassignRegionHandler(124): Close a739dece698030682185693f1769a5eb 2025-01-08T05:01:14,653 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T05:01:14,654 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] regionserver.HRegion(1681): Closing a739dece698030682185693f1769a5eb, disabling compactions & flushes 2025-01-08T05:01:14,654 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452763.a739dece698030682185693f1769a5eb. 2025-01-08T05:01:14,654 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452763.a739dece698030682185693f1769a5eb. 2025-01-08T05:01:14,654 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452763.a739dece698030682185693f1769a5eb. after waiting 0 ms 2025-01-08T05:01:14,654 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452763.a739dece698030682185693f1769a5eb. 2025-01-08T05:01:14,659 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a739dece698030682185693f1769a5eb/recovered.edits/12.seqid, newMaxSeqId=12, maxSeqId=8 2025-01-08T05:01:14,659 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:01:14,660 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452763.a739dece698030682185693f1769a5eb. 2025-01-08T05:01:14,660 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] regionserver.HRegion(1635): Region close journal for a739dece698030682185693f1769a5eb: 2025-01-08T05:01:14,661 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] handler.UnassignRegionHandler(170): Closed a739dece698030682185693f1769a5eb 2025-01-08T05:01:14,662 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=126 updating hbase:meta row=a739dece698030682185693f1769a5eb, regionState=CLOSED 2025-01-08T05:01:14,665 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=127, resume processing ppid=126 2025-01-08T05:01:14,665 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=127, ppid=126, state=SUCCESS; CloseRegionProcedure a739dece698030682185693f1769a5eb, server=f28094b72178,37967,1736312288941 in 162 msec 2025-01-08T05:01:14,666 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=126, resume processing ppid=125 2025-01-08T05:01:14,666 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=126, ppid=125, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=a739dece698030682185693f1769a5eb, UNASSIGN in 166 msec 2025-01-08T05:01:14,668 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=125, resume processing ppid=124 2025-01-08T05:01:14,668 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=125, ppid=124, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 in 170 msec 2025-01-08T05:01:14,670 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion-1","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312474669"}]},"ts":"1736312474669"} 2025-01-08T05:01:14,671 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithMergeRegion-1, state=DISABLED in hbase:meta 2025-01-08T05:01:14,674 INFO [PEWorker-5 {}] procedure.DisableTableProcedure(296): Set testtb-testExportFileSystemStateWithMergeRegion-1 to state=DISABLED 2025-01-08T05:01:14,677 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=124, state=SUCCESS; DisableTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 in 185 msec 2025-01-08T05:01:14,794 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=124 2025-01-08T05:01:14,795 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion-1, procId: 124 completed 2025-01-08T05:01:14,795 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.3 delete testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:01:14,796 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=128, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:01:14,797 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=128, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:01:14,797 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:01:14,798 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=128, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:01:14,799 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37583 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:01:14,800 DEBUG [HFileArchiver-11 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a739dece698030682185693f1769a5eb 2025-01-08T05:01:14,800 DEBUG [HFileArchiver-12 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/4e1ff0e98851e58decfd4b7963bd4305 2025-01-08T05:01:14,801 DEBUG [HFileArchiver-13 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/747352f4d6112e9ff87f40579cee217d 2025-01-08T05:01:14,802 DEBUG [HFileArchiver-13 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/747352f4d6112e9ff87f40579cee217d/cf, FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/747352f4d6112e9ff87f40579cee217d/recovered.edits] 2025-01-08T05:01:14,802 DEBUG [HFileArchiver-11 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a739dece698030682185693f1769a5eb/cf, FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a739dece698030682185693f1769a5eb/recovered.edits] 2025-01-08T05:01:14,802 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:01:14,802 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:01:14,803 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:01:14,803 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:01:14,804 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion-1 with data PBUF 2025-01-08T05:01:14,804 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion-1 with data PBUF 2025-01-08T05:01:14,804 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion-1 with data PBUF 2025-01-08T05:01:14,804 DEBUG [HFileArchiver-12 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/4e1ff0e98851e58decfd4b7963bd4305/cf, FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/4e1ff0e98851e58decfd4b7963bd4305/recovered.edits] 2025-01-08T05:01:14,805 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:01:14,805 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:01:14,805 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:14,805 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:14,805 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:14,806 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:01:14,806 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:14,806 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=128 2025-01-08T05:01:14,807 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:14,807 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:14,807 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:14,809 DEBUG [HFileArchiver-11 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a739dece698030682185693f1769a5eb/cf/594ade9bfa6147828e69c650755e989f.747352f4d6112e9ff87f40579cee217d to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a739dece698030682185693f1769a5eb/cf/594ade9bfa6147828e69c650755e989f.747352f4d6112e9ff87f40579cee217d 2025-01-08T05:01:14,810 DEBUG [HFileArchiver-12 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/4e1ff0e98851e58decfd4b7963bd4305/cf/e7b558af0b6c41c7a0cc026545a9e6bb to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/4e1ff0e98851e58decfd4b7963bd4305/cf/e7b558af0b6c41c7a0cc026545a9e6bb 2025-01-08T05:01:14,811 DEBUG [HFileArchiver-11 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a739dece698030682185693f1769a5eb/cf/e7b558af0b6c41c7a0cc026545a9e6bb.4e1ff0e98851e58decfd4b7963bd4305 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a739dece698030682185693f1769a5eb/cf/e7b558af0b6c41c7a0cc026545a9e6bb.4e1ff0e98851e58decfd4b7963bd4305 2025-01-08T05:01:14,812 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion-1 with data null 2025-01-08T05:01:14,812 INFO [zk-permission-watcher-pool-0 {}] access.AuthManager(135): Skipping permission cache refresh because writable data is empty 2025-01-08T05:01:14,813 DEBUG [HFileArchiver-13 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/747352f4d6112e9ff87f40579cee217d/cf/594ade9bfa6147828e69c650755e989f to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/747352f4d6112e9ff87f40579cee217d/cf/594ade9bfa6147828e69c650755e989f 2025-01-08T05:01:14,814 DEBUG [HFileArchiver-12 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/4e1ff0e98851e58decfd4b7963bd4305/recovered.edits/8.seqid to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/4e1ff0e98851e58decfd4b7963bd4305/recovered.edits/8.seqid 2025-01-08T05:01:14,814 DEBUG [HFileArchiver-11 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a739dece698030682185693f1769a5eb/recovered.edits/12.seqid to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a739dece698030682185693f1769a5eb/recovered.edits/12.seqid 2025-01-08T05:01:14,815 DEBUG [HFileArchiver-12 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/4e1ff0e98851e58decfd4b7963bd4305 2025-01-08T05:01:14,815 DEBUG [HFileArchiver-11 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a739dece698030682185693f1769a5eb 2025-01-08T05:01:14,816 DEBUG [HFileArchiver-13 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/747352f4d6112e9ff87f40579cee217d/recovered.edits/8.seqid to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/747352f4d6112e9ff87f40579cee217d/recovered.edits/8.seqid 2025-01-08T05:01:14,817 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:14,817 DEBUG [HFileArchiver-13 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/747352f4d6112e9ff87f40579cee217d 2025-01-08T05:01:14,817 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(313): Archived testtb-testExportFileSystemStateWithMergeRegion-1 regions 2025-01-08T05:01:14,819 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=128, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:01:14,821 WARN [PEWorker-1 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 1 rows of testtb-testExportFileSystemStateWithMergeRegion-1 from hbase:meta 2025-01-08T05:01:14,823 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testExportFileSystemStateWithMergeRegion-1' descriptor. 2025-01-08T05:01:14,824 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=128, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:01:14,824 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testExportFileSystemStateWithMergeRegion-1' from region states. 2025-01-08T05:01:14,824 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452763.a739dece698030682185693f1769a5eb.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736312474824"}]},"ts":"9223372036854775807"} 2025-01-08T05:01:14,825 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1808): Deleted 1 regions from META 2025-01-08T05:01:14,826 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => a739dece698030682185693f1769a5eb, NAME => 'testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452763.a739dece698030682185693f1769a5eb.', STARTKEY => '', ENDKEY => ''}] 2025-01-08T05:01:14,826 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testExportFileSystemStateWithMergeRegion-1' as deleted. 2025-01-08T05:01:14,826 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion-1","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736312474826"}]},"ts":"9223372036854775807"} 2025-01-08T05:01:14,827 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testExportFileSystemStateWithMergeRegion-1 state from META 2025-01-08T05:01:14,829 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(133): Finished pid=128, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:01:14,830 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=128, state=SUCCESS; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 in 33 msec 2025-01-08T05:01:14,907 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=128 2025-01-08T05:01:14,907 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion-1, procId: 128 completed 2025-01-08T05:01:14,908 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T05:01:14,908 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.3 disable testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T05:01:14,909 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=129, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T05:01:14,910 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=129 2025-01-08T05:01:14,911 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312474911"}]},"ts":"1736312474911"} 2025-01-08T05:01:14,912 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithMergeRegion, state=DISABLING in hbase:meta 2025-01-08T05:01:14,919 INFO [PEWorker-4 {}] procedure.DisableTableProcedure(284): Set testtb-testExportFileSystemStateWithMergeRegion to state=DISABLING 2025-01-08T05:01:14,919 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=130, ppid=129, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testExportFileSystemStateWithMergeRegion}] 2025-01-08T05:01:14,921 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=131, ppid=130, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=762016803bad87dc082747011636b848, UNASSIGN}, {pid=132, ppid=130, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=8ceb1f1fe1c86dfb459738a7200e1333, UNASSIGN}] 2025-01-08T05:01:14,921 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=132, ppid=130, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=8ceb1f1fe1c86dfb459738a7200e1333, UNASSIGN 2025-01-08T05:01:14,921 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=131, ppid=130, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=762016803bad87dc082747011636b848, UNASSIGN 2025-01-08T05:01:14,922 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=132 updating hbase:meta row=8ceb1f1fe1c86dfb459738a7200e1333, regionState=CLOSING, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T05:01:14,922 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=131 updating hbase:meta row=762016803bad87dc082747011636b848, regionState=CLOSING, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T05:01:14,923 DEBUG [PEWorker-5 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T05:01:14,923 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=133, ppid=131, state=RUNNABLE; CloseRegionProcedure 762016803bad87dc082747011636b848, server=f28094b72178,37967,1736312288941}] 2025-01-08T05:01:14,924 DEBUG [PEWorker-2 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T05:01:14,924 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=134, ppid=132, state=RUNNABLE; CloseRegionProcedure 8ceb1f1fe1c86dfb459738a7200e1333, server=f28094b72178,37583,1736312288873}] 2025-01-08T05:01:15,011 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=129 2025-01-08T05:01:15,075 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:01:15,076 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] handler.UnassignRegionHandler(124): Close 762016803bad87dc082747011636b848 2025-01-08T05:01:15,076 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T05:01:15,076 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] regionserver.HRegion(1681): Closing 762016803bad87dc082747011636b848, disabling compactions & flushes 2025-01-08T05:01:15,076 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848. 2025-01-08T05:01:15,076 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848. 2025-01-08T05:01:15,076 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848. after waiting 0 ms 2025-01-08T05:01:15,076 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848. 2025-01-08T05:01:15,076 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T05:01:15,076 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] handler.UnassignRegionHandler(124): Close 8ceb1f1fe1c86dfb459738a7200e1333 2025-01-08T05:01:15,076 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T05:01:15,077 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] regionserver.HRegion(1681): Closing 8ceb1f1fe1c86dfb459738a7200e1333, disabling compactions & flushes 2025-01-08T05:01:15,077 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333. 2025-01-08T05:01:15,077 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333. 2025-01-08T05:01:15,077 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333. after waiting 0 ms 2025-01-08T05:01:15,077 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333. 2025-01-08T05:01:15,080 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/762016803bad87dc082747011636b848/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T05:01:15,080 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/8ceb1f1fe1c86dfb459738a7200e1333/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T05:01:15,081 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:01:15,081 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:01:15,081 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848. 2025-01-08T05:01:15,081 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] regionserver.HRegion(1635): Region close journal for 762016803bad87dc082747011636b848: 2025-01-08T05:01:15,081 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333. 2025-01-08T05:01:15,081 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] regionserver.HRegion(1635): Region close journal for 8ceb1f1fe1c86dfb459738a7200e1333: 2025-01-08T05:01:15,082 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] handler.UnassignRegionHandler(170): Closed 8ceb1f1fe1c86dfb459738a7200e1333 2025-01-08T05:01:15,083 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=132 updating hbase:meta row=8ceb1f1fe1c86dfb459738a7200e1333, regionState=CLOSED 2025-01-08T05:01:15,083 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] handler.UnassignRegionHandler(170): Closed 762016803bad87dc082747011636b848 2025-01-08T05:01:15,083 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=131 updating hbase:meta row=762016803bad87dc082747011636b848, regionState=CLOSED 2025-01-08T05:01:15,088 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=134, resume processing ppid=132 2025-01-08T05:01:15,088 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=134, ppid=132, state=SUCCESS; CloseRegionProcedure 8ceb1f1fe1c86dfb459738a7200e1333, server=f28094b72178,37583,1736312288873 in 163 msec 2025-01-08T05:01:15,089 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=133, resume processing ppid=131 2025-01-08T05:01:15,089 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=133, ppid=131, state=SUCCESS; CloseRegionProcedure 762016803bad87dc082747011636b848, server=f28094b72178,37967,1736312288941 in 164 msec 2025-01-08T05:01:15,089 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=132, ppid=130, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=8ceb1f1fe1c86dfb459738a7200e1333, UNASSIGN in 167 msec 2025-01-08T05:01:15,090 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=131, resume processing ppid=130 2025-01-08T05:01:15,090 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=131, ppid=130, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=762016803bad87dc082747011636b848, UNASSIGN in 168 msec 2025-01-08T05:01:15,092 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=130, resume processing ppid=129 2025-01-08T05:01:15,092 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=130, ppid=129, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testExportFileSystemStateWithMergeRegion in 171 msec 2025-01-08T05:01:15,093 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312475093"}]},"ts":"1736312475093"} 2025-01-08T05:01:15,101 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithMergeRegion, state=DISABLED in hbase:meta 2025-01-08T05:01:15,103 INFO [PEWorker-3 {}] procedure.DisableTableProcedure(296): Set testtb-testExportFileSystemStateWithMergeRegion to state=DISABLED 2025-01-08T05:01:15,105 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=129, state=SUCCESS; DisableTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion in 196 msec 2025-01-08T05:01:15,212 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=129 2025-01-08T05:01:15,213 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion, procId: 129 completed 2025-01-08T05:01:15,213 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.3 delete testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T05:01:15,227 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=135, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T05:01:15,230 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=135, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T05:01:15,230 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T05:01:15,232 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=135, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T05:01:15,233 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37583 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T05:01:15,235 DEBUG [HFileArchiver-14 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/762016803bad87dc082747011636b848 2025-01-08T05:01:15,235 DEBUG [HFileArchiver-15 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/8ceb1f1fe1c86dfb459738a7200e1333 2025-01-08T05:01:15,237 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T05:01:15,237 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T05:01:15,239 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T05:01:15,239 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF 2025-01-08T05:01:15,239 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF 2025-01-08T05:01:15,241 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T05:01:15,241 DEBUG [HFileArchiver-14 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/762016803bad87dc082747011636b848/cf, FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/762016803bad87dc082747011636b848/recovered.edits] 2025-01-08T05:01:15,241 DEBUG [HFileArchiver-15 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/8ceb1f1fe1c86dfb459738a7200e1333/cf, FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/8ceb1f1fe1c86dfb459738a7200e1333/recovered.edits] 2025-01-08T05:01:15,241 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF 2025-01-08T05:01:15,242 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF 2025-01-08T05:01:15,243 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T05:01:15,243 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T05:01:15,243 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:15,243 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:15,243 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T05:01:15,243 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:15,243 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T05:01:15,243 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:15,244 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=135 2025-01-08T05:01:15,246 DEBUG [HFileArchiver-14 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/762016803bad87dc082747011636b848/cf/0b3ad9b3094546f082040c46a082079c to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion/762016803bad87dc082747011636b848/cf/0b3ad9b3094546f082040c46a082079c 2025-01-08T05:01:15,246 DEBUG [HFileArchiver-15 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/8ceb1f1fe1c86dfb459738a7200e1333/cf/ad6eb7d3730a4aa682e10bc5d132a8ee to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion/8ceb1f1fe1c86dfb459738a7200e1333/cf/ad6eb7d3730a4aa682e10bc5d132a8ee 2025-01-08T05:01:15,249 DEBUG [HFileArchiver-14 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/762016803bad87dc082747011636b848/recovered.edits/9.seqid to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion/762016803bad87dc082747011636b848/recovered.edits/9.seqid 2025-01-08T05:01:15,250 DEBUG [HFileArchiver-15 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/8ceb1f1fe1c86dfb459738a7200e1333/recovered.edits/9.seqid to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion/8ceb1f1fe1c86dfb459738a7200e1333/recovered.edits/9.seqid 2025-01-08T05:01:15,250 DEBUG [HFileArchiver-14 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/762016803bad87dc082747011636b848 2025-01-08T05:01:15,253 DEBUG [HFileArchiver-15 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithMergeRegion/8ceb1f1fe1c86dfb459738a7200e1333 2025-01-08T05:01:15,253 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(313): Archived testtb-testExportFileSystemStateWithMergeRegion regions 2025-01-08T05:01:15,254 DEBUG [PEWorker-1 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c 2025-01-08T05:01:15,255 DEBUG [PEWorker-1 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf] 2025-01-08T05:01:15,260 DEBUG [PEWorker-1 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/c4ca4238a0b923820dcc509a6f75849b2025010845a19cd429654dad94fb8e4d1c4f5cc4_8ceb1f1fe1c86dfb459738a7200e1333 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/c4ca4238a0b923820dcc509a6f75849b2025010845a19cd429654dad94fb8e4d1c4f5cc4_8ceb1f1fe1c86dfb459738a7200e1333 2025-01-08T05:01:15,261 DEBUG [PEWorker-1 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/d41d8cd98f00b204e9800998ecf8427e202501087e1ddbc52f9a4cba91dd7aa7111f6827_762016803bad87dc082747011636b848 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/d41d8cd98f00b204e9800998ecf8427e202501087e1ddbc52f9a4cba91dd7aa7111f6827_762016803bad87dc082747011636b848 2025-01-08T05:01:15,262 DEBUG [PEWorker-1 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c 2025-01-08T05:01:15,265 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=135, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T05:01:15,267 WARN [PEWorker-1 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testtb-testExportFileSystemStateWithMergeRegion from hbase:meta 2025-01-08T05:01:15,270 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testExportFileSystemStateWithMergeRegion' descriptor. 2025-01-08T05:01:15,271 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=135, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T05:01:15,271 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testExportFileSystemStateWithMergeRegion' from region states. 2025-01-08T05:01:15,271 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736312475271"}]},"ts":"9223372036854775807"} 2025-01-08T05:01:15,271 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736312475271"}]},"ts":"9223372036854775807"} 2025-01-08T05:01:15,273 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T05:01:15,273 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => 762016803bad87dc082747011636b848, NAME => 'testtb-testExportFileSystemStateWithMergeRegion,,1736312451439.762016803bad87dc082747011636b848.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => 8ceb1f1fe1c86dfb459738a7200e1333, NAME => 'testtb-testExportFileSystemStateWithMergeRegion,1,1736312451439.8ceb1f1fe1c86dfb459738a7200e1333.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T05:01:15,273 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testExportFileSystemStateWithMergeRegion' as deleted. 2025-01-08T05:01:15,273 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736312475273"}]},"ts":"9223372036854775807"} 2025-01-08T05:01:15,276 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testExportFileSystemStateWithMergeRegion state from META 2025-01-08T05:01:15,278 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(133): Finished pid=135, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T05:01:15,279 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=135, state=SUCCESS; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion in 65 msec 2025-01-08T05:01:15,345 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=135 2025-01-08T05:01:15,345 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion, procId: 135 completed 2025-01-08T05:01:15,355 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "emptySnaptb0-testExportFileSystemStateWithMergeRegion" 2025-01-08T05:01:15,357 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(380): Deleting snapshot: emptySnaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T05:01:15,359 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "snaptb0-testExportFileSystemStateWithMergeRegion" 2025-01-08T05:01:15,360 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T05:01:15,362 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "snaptb0-testExportFileSystemStateWithMergeRegion-1" 2025-01-08T05:01:15,364 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:01:15,395 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: snapshot.TestMobSecureExportSnapshot#testExportFileSystemStateWithMergeRegion Thread=799 (was 787) Potentially hanging thread: IPC Client (366384994) connection to localhost/127.0.0.1:45809 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-575728404_1 at /127.0.0.1:44870 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-11 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Thread-4972 java.base@17.0.11/java.io.FileInputStream.readBytes(Native Method) java.base@17.0.11/java.io.FileInputStream.read(FileInputStream.java:276) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:282) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) java.base@17.0.11/sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:281) java.base@17.0.11/sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:324) java.base@17.0.11/sun.nio.cs.StreamDecoder.read(StreamDecoder.java:189) java.base@17.0.11/java.io.InputStreamReader.read(InputStreamReader.java:177) java.base@17.0.11/java.io.BufferedReader.fill(BufferedReader.java:162) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:329) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:396) app//org.apache.hadoop.util.Shell$1.run(Shell.java:1025) Potentially hanging thread: hconnection-0x168697cd-shared-pool-28 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x168697cd-shared-pool-31 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1149249371_22 at /127.0.0.1:39024 [Waiting for operation #4] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x168697cd-shared-pool-29 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1149249371_22 at /127.0.0.1:38878 [Waiting for operation #5] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #10 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x168697cd-shared-pool-30 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-6-5 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:62) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:883) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:45367 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-6-6 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:62) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:883) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1149249371_22 at /127.0.0.1:39806 [Waiting for operation #5] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-13 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-575728404_1 at /127.0.0.1:54420 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-6-7 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:62) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:883) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: process reaper (pid 21200) java.base@17.0.11/java.lang.ProcessHandleImpl.waitForProcessExit0(Native Method) java.base@17.0.11/java.lang.ProcessHandleImpl$1.run(ProcessHandleImpl.java:150) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x168697cd-shared-pool-32 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #9 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-15 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x168697cd-shared-pool-33 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (366384994) connection to localhost/127.0.0.1:45367 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: HFileArchiver-14 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-12 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) - Thread LEAK? -, OpenFileDescriptor=813 (was 799) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=952 (was 1164), ProcessCount=14 (was 17), AvailableMemoryMB=2450 (was 2712) 2025-01-08T05:01:15,396 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=799 is superior to 500 2025-01-08T05:01:15,416 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: snapshot.TestMobSecureExportSnapshot#testExportExpiredSnapshot Thread=799, OpenFileDescriptor=813, MaxFileDescriptor=1048576, SystemLoadAverage=952, ProcessCount=14, AvailableMemoryMB=2441 2025-01-08T05:01:15,416 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=799 is superior to 500 2025-01-08T05:01:15,418 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.3 create 'testtb-testExportExpiredSnapshot', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T05:01:15,419 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=136, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testExportExpiredSnapshot 2025-01-08T05:01:15,420 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=136, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportExpiredSnapshot execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T05:01:15,421 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "testtb-testExportExpiredSnapshot" procId is: 136 2025-01-08T05:01:15,422 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=136 2025-01-08T05:01:15,422 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=136, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testExportExpiredSnapshot execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T05:01:15,430 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742193_1369 (size=443) 2025-01-08T05:01:15,431 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742193_1369 (size=443) 2025-01-08T05:01:15,431 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742193_1369 (size=443) 2025-01-08T05:01:15,433 INFO [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 4718f4a49bc4b9e61ea5fdb2dbd018ec, NAME => 'testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testtb-testExportExpiredSnapshot', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:01:15,434 INFO [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => 8ddd1acb9e69eae44cba4b9604557fed, NAME => 'testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testtb-testExportExpiredSnapshot', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:01:15,447 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742195_1371 (size=68) 2025-01-08T05:01:15,447 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742195_1371 (size=68) 2025-01-08T05:01:15,447 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742195_1371 (size=68) 2025-01-08T05:01:15,448 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742194_1370 (size=68) 2025-01-08T05:01:15,448 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:01:15,449 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1681): Closing 8ddd1acb9e69eae44cba4b9604557fed, disabling compactions & flushes 2025-01-08T05:01:15,449 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742194_1370 (size=68) 2025-01-08T05:01:15,449 INFO [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed. 2025-01-08T05:01:15,449 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed. 2025-01-08T05:01:15,449 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed. after waiting 0 ms 2025-01-08T05:01:15,449 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed. 2025-01-08T05:01:15,449 INFO [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed. 2025-01-08T05:01:15,449 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1635): Region close journal for 8ddd1acb9e69eae44cba4b9604557fed: 2025-01-08T05:01:15,449 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742194_1370 (size=68) 2025-01-08T05:01:15,450 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:01:15,450 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1681): Closing 4718f4a49bc4b9e61ea5fdb2dbd018ec, disabling compactions & flushes 2025-01-08T05:01:15,450 INFO [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec. 2025-01-08T05:01:15,450 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec. 2025-01-08T05:01:15,450 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec. after waiting 0 ms 2025-01-08T05:01:15,450 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec. 2025-01-08T05:01:15,450 INFO [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec. 2025-01-08T05:01:15,450 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1635): Region close journal for 4718f4a49bc4b9e61ea5fdb2dbd018ec: 2025-01-08T05:01:15,451 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=136, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testExportExpiredSnapshot execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T05:01:15,451 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed.","families":{"info":[{"qualifier":"regioninfo","vlen":67,"tag":[],"timestamp":"1736312475451"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736312475451"}]},"ts":"1736312475451"} 2025-01-08T05:01:15,451 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec.","families":{"info":[{"qualifier":"regioninfo","vlen":67,"tag":[],"timestamp":"1736312475451"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736312475451"}]},"ts":"1736312475451"} 2025-01-08T05:01:15,453 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T05:01:15,455 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=136, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testExportExpiredSnapshot execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T05:01:15,455 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportExpiredSnapshot","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312475455"}]},"ts":"1736312475455"} 2025-01-08T05:01:15,456 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportExpiredSnapshot, state=ENABLING in hbase:meta 2025-01-08T05:01:15,460 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(202): Hosts are {f28094b72178=0} racks are {/default-rack=0} 2025-01-08T05:01:15,462 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T05:01:15,462 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T05:01:15,462 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T05:01:15,462 INFO [PEWorker-2 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T05:01:15,462 INFO [PEWorker-2 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T05:01:15,462 INFO [PEWorker-2 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T05:01:15,462 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T05:01:15,462 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=137, ppid=136, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=4718f4a49bc4b9e61ea5fdb2dbd018ec, ASSIGN}, {pid=138, ppid=136, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=8ddd1acb9e69eae44cba4b9604557fed, ASSIGN}] 2025-01-08T05:01:15,464 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=138, ppid=136, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=8ddd1acb9e69eae44cba4b9604557fed, ASSIGN 2025-01-08T05:01:15,464 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=137, ppid=136, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=4718f4a49bc4b9e61ea5fdb2dbd018ec, ASSIGN 2025-01-08T05:01:15,465 INFO [PEWorker-5 {}] assignment.TransitRegionStateProcedure(264): Starting pid=138, ppid=136, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=8ddd1acb9e69eae44cba4b9604557fed, ASSIGN; state=OFFLINE, location=f28094b72178,39801,1736312288779; forceNewPlan=false, retain=false 2025-01-08T05:01:15,465 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=137, ppid=136, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=4718f4a49bc4b9e61ea5fdb2dbd018ec, ASSIGN; state=OFFLINE, location=f28094b72178,37967,1736312288941; forceNewPlan=false, retain=false 2025-01-08T05:01:15,523 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=136 2025-01-08T05:01:15,615 INFO [f28094b72178:42195 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T05:01:15,615 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=138 updating hbase:meta row=8ddd1acb9e69eae44cba4b9604557fed, regionState=OPENING, regionLocation=f28094b72178,39801,1736312288779 2025-01-08T05:01:15,615 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=137 updating hbase:meta row=4718f4a49bc4b9e61ea5fdb2dbd018ec, regionState=OPENING, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T05:01:15,617 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=139, ppid=138, state=RUNNABLE; OpenRegionProcedure 8ddd1acb9e69eae44cba4b9604557fed, server=f28094b72178,39801,1736312288779}] 2025-01-08T05:01:15,618 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=140, ppid=137, state=RUNNABLE; OpenRegionProcedure 4718f4a49bc4b9e61ea5fdb2dbd018ec, server=f28094b72178,37967,1736312288941}] 2025-01-08T05:01:15,724 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=136 2025-01-08T05:01:15,768 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,39801,1736312288779 2025-01-08T05:01:15,770 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:01:15,772 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] handler.AssignRegionHandler(135): Open testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed. 2025-01-08T05:01:15,772 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(7285): Opening region: {ENCODED => 8ddd1acb9e69eae44cba4b9604557fed, NAME => 'testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T05:01:15,772 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed. service=AccessControlService 2025-01-08T05:01:15,773 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] handler.AssignRegionHandler(135): Open testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec. 2025-01-08T05:01:15,773 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T05:01:15,773 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(7285): Opening region: {ENCODED => 4718f4a49bc4b9e61ea5fdb2dbd018ec, NAME => 'testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T05:01:15,773 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportExpiredSnapshot 8ddd1acb9e69eae44cba4b9604557fed 2025-01-08T05:01:15,773 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(894): Instantiated testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:01:15,773 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(7327): checking encryption for 8ddd1acb9e69eae44cba4b9604557fed 2025-01-08T05:01:15,773 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(7330): checking classloading for 8ddd1acb9e69eae44cba4b9604557fed 2025-01-08T05:01:15,773 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec. service=AccessControlService 2025-01-08T05:01:15,773 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T05:01:15,773 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportExpiredSnapshot 4718f4a49bc4b9e61ea5fdb2dbd018ec 2025-01-08T05:01:15,773 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(894): Instantiated testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:01:15,773 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(7327): checking encryption for 4718f4a49bc4b9e61ea5fdb2dbd018ec 2025-01-08T05:01:15,774 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(7330): checking classloading for 4718f4a49bc4b9e61ea5fdb2dbd018ec 2025-01-08T05:01:15,775 INFO [StoreOpener-8ddd1acb9e69eae44cba4b9604557fed-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 8ddd1acb9e69eae44cba4b9604557fed 2025-01-08T05:01:15,775 INFO [StoreOpener-4718f4a49bc4b9e61ea5fdb2dbd018ec-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 4718f4a49bc4b9e61ea5fdb2dbd018ec 2025-01-08T05:01:15,776 INFO [StoreOpener-8ddd1acb9e69eae44cba4b9604557fed-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 8ddd1acb9e69eae44cba4b9604557fed columnFamilyName cf 2025-01-08T05:01:15,776 INFO [StoreOpener-4718f4a49bc4b9e61ea5fdb2dbd018ec-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 4718f4a49bc4b9e61ea5fdb2dbd018ec columnFamilyName cf 2025-01-08T05:01:15,777 DEBUG [StoreOpener-8ddd1acb9e69eae44cba4b9604557fed-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:01:15,778 DEBUG [StoreOpener-4718f4a49bc4b9e61ea5fdb2dbd018ec-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:01:15,778 INFO [StoreOpener-8ddd1acb9e69eae44cba4b9604557fed-1 {}] regionserver.HStore(327): Store=8ddd1acb9e69eae44cba4b9604557fed/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T05:01:15,779 INFO [StoreOpener-4718f4a49bc4b9e61ea5fdb2dbd018ec-1 {}] regionserver.HStore(327): Store=4718f4a49bc4b9e61ea5fdb2dbd018ec/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T05:01:15,780 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/8ddd1acb9e69eae44cba4b9604557fed 2025-01-08T05:01:15,780 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/4718f4a49bc4b9e61ea5fdb2dbd018ec 2025-01-08T05:01:15,780 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/8ddd1acb9e69eae44cba4b9604557fed 2025-01-08T05:01:15,781 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/4718f4a49bc4b9e61ea5fdb2dbd018ec 2025-01-08T05:01:15,783 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(1085): writing seq id for 4718f4a49bc4b9e61ea5fdb2dbd018ec 2025-01-08T05:01:15,783 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(1085): writing seq id for 8ddd1acb9e69eae44cba4b9604557fed 2025-01-08T05:01:15,785 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/4718f4a49bc4b9e61ea5fdb2dbd018ec/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T05:01:15,785 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(1102): Opened 4718f4a49bc4b9e61ea5fdb2dbd018ec; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=67307735, jitterRate=0.0029634088277816772}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T05:01:15,786 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(1001): Region open journal for 4718f4a49bc4b9e61ea5fdb2dbd018ec: 2025-01-08T05:01:15,787 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec., pid=140, masterSystemTime=1736312475769 2025-01-08T05:01:15,787 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/8ddd1acb9e69eae44cba4b9604557fed/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T05:01:15,788 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(1102): Opened 8ddd1acb9e69eae44cba4b9604557fed; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=73703796, jitterRate=0.09827214479446411}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T05:01:15,788 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(1001): Region open journal for 8ddd1acb9e69eae44cba4b9604557fed: 2025-01-08T05:01:15,788 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec. 2025-01-08T05:01:15,788 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] handler.AssignRegionHandler(164): Opened testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec. 2025-01-08T05:01:15,788 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed., pid=139, masterSystemTime=1736312475768 2025-01-08T05:01:15,789 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=137 updating hbase:meta row=4718f4a49bc4b9e61ea5fdb2dbd018ec, regionState=OPEN, openSeqNum=2, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T05:01:15,790 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed. 2025-01-08T05:01:15,790 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] handler.AssignRegionHandler(164): Opened testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed. 2025-01-08T05:01:15,790 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=138 updating hbase:meta row=8ddd1acb9e69eae44cba4b9604557fed, regionState=OPEN, openSeqNum=2, regionLocation=f28094b72178,39801,1736312288779 2025-01-08T05:01:15,792 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=140, resume processing ppid=137 2025-01-08T05:01:15,792 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=140, ppid=137, state=SUCCESS; OpenRegionProcedure 4718f4a49bc4b9e61ea5fdb2dbd018ec, server=f28094b72178,37967,1736312288941 in 173 msec 2025-01-08T05:01:15,793 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=139, resume processing ppid=138 2025-01-08T05:01:15,793 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=139, ppid=138, state=SUCCESS; OpenRegionProcedure 8ddd1acb9e69eae44cba4b9604557fed, server=f28094b72178,39801,1736312288779 in 174 msec 2025-01-08T05:01:15,793 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=137, ppid=136, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=4718f4a49bc4b9e61ea5fdb2dbd018ec, ASSIGN in 330 msec 2025-01-08T05:01:15,795 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=138, resume processing ppid=136 2025-01-08T05:01:15,795 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=138, ppid=136, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=8ddd1acb9e69eae44cba4b9604557fed, ASSIGN in 331 msec 2025-01-08T05:01:15,795 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=136, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testExportExpiredSnapshot execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T05:01:15,796 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportExpiredSnapshot","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312475795"}]},"ts":"1736312475795"} 2025-01-08T05:01:15,797 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportExpiredSnapshot, state=ENABLED in hbase:meta 2025-01-08T05:01:15,799 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=136, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportExpiredSnapshot execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T05:01:15,799 DEBUG [PEWorker-5 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testExportExpiredSnapshot jenkins: RWXCA 2025-01-08T05:01:15,801 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37583 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportExpiredSnapshot], kv [jenkins: RWXCA] 2025-01-08T05:01:15,803 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:15,803 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:15,803 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:15,803 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:15,805 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:15,805 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:15,805 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:15,805 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:15,807 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=136, state=SUCCESS; CreateTableProcedure table=testtb-testExportExpiredSnapshot in 387 msec 2025-01-08T05:01:16,025 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=136 2025-01-08T05:01:16,025 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testExportExpiredSnapshot, procId: 136 completed 2025-01-08T05:01:16,027 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportExpiredSnapshot 2025-01-08T05:01:16,027 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec. 2025-01-08T05:01:16,028 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T05:01:16,040 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } 2025-01-08T05:01:16,040 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736312476040 (current time:1736312476040). 2025-01-08T05:01:16,040 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T05:01:16,040 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot emptySnaptb0-testExportExpiredSnapshot VERSION not specified, setting to 2 2025-01-08T05:01:16,040 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T05:01:16,041 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x3fdd9fd8 to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@24023d89 2025-01-08T05:01:16,050 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@46234ccc, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:01:16,052 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:01:16,053 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:43184, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:01:16,054 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x3fdd9fd8 to 127.0.0.1:59048 2025-01-08T05:01:16,054 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:01:16,055 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x44fcce9f to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@71274dfc 2025-01-08T05:01:16,063 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2d3e3a38, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:01:16,065 DEBUG [hconnection-0x5f474c9e-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:01:16,066 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:43190, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:01:16,068 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x44fcce9f to 127.0.0.1:59048 2025-01-08T05:01:16,068 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:01:16,068 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportExpiredSnapshot], kv [jenkins: RWXCA] 2025-01-08T05:01:16,069 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T05:01:16,070 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=141, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } 2025-01-08T05:01:16,070 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 }, snapshot procedure id = 141 2025-01-08T05:01:16,070 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T05:01:16,071 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=141 2025-01-08T05:01:16,073 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T05:01:16,075 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T05:01:16,091 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742196_1372 (size=170) 2025-01-08T05:01:16,091 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742196_1372 (size=170) 2025-01-08T05:01:16,091 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742196_1372 (size=170) 2025-01-08T05:01:16,092 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T05:01:16,093 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=142, ppid=141, state=RUNNABLE; SnapshotRegionProcedure 4718f4a49bc4b9e61ea5fdb2dbd018ec}, {pid=143, ppid=141, state=RUNNABLE; SnapshotRegionProcedure 8ddd1acb9e69eae44cba4b9604557fed}] 2025-01-08T05:01:16,093 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=142, ppid=141, state=RUNNABLE; SnapshotRegionProcedure 4718f4a49bc4b9e61ea5fdb2dbd018ec 2025-01-08T05:01:16,094 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=143, ppid=141, state=RUNNABLE; SnapshotRegionProcedure 8ddd1acb9e69eae44cba4b9604557fed 2025-01-08T05:01:16,172 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=141 2025-01-08T05:01:16,245 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:01:16,245 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,39801,1736312288779 2025-01-08T05:01:16,246 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=39801 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=143 2025-01-08T05:01:16,246 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=37967 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=142 2025-01-08T05:01:16,246 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=142}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec. 2025-01-08T05:01:16,246 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=143}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed. 2025-01-08T05:01:16,246 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=143}] regionserver.HRegion(2538): Flush status journal for 8ddd1acb9e69eae44cba4b9604557fed: 2025-01-08T05:01:16,246 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=142}] regionserver.HRegion(2538): Flush status journal for 4718f4a49bc4b9e61ea5fdb2dbd018ec: 2025-01-08T05:01:16,246 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=143}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed. for emptySnaptb0-testExportExpiredSnapshot completed. 2025-01-08T05:01:16,246 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=142}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec. for emptySnaptb0-testExportExpiredSnapshot completed. 2025-01-08T05:01:16,246 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=142}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec.' region-info for snapshot=emptySnaptb0-testExportExpiredSnapshot 2025-01-08T05:01:16,246 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=143}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed.' region-info for snapshot=emptySnaptb0-testExportExpiredSnapshot 2025-01-08T05:01:16,246 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=143}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T05:01:16,246 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=142}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T05:01:16,246 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=142}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T05:01:16,246 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=143}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T05:01:16,254 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742197_1373 (size=71) 2025-01-08T05:01:16,255 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742197_1373 (size=71) 2025-01-08T05:01:16,255 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742197_1373 (size=71) 2025-01-08T05:01:16,255 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=142}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec. 2025-01-08T05:01:16,255 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=142}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=142 2025-01-08T05:01:16,256 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=142 2025-01-08T05:01:16,256 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportExpiredSnapshot on region 4718f4a49bc4b9e61ea5fdb2dbd018ec 2025-01-08T05:01:16,256 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=142, ppid=141, state=RUNNABLE; SnapshotRegionProcedure 4718f4a49bc4b9e61ea5fdb2dbd018ec 2025-01-08T05:01:16,259 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=142, ppid=141, state=SUCCESS; SnapshotRegionProcedure 4718f4a49bc4b9e61ea5fdb2dbd018ec in 164 msec 2025-01-08T05:01:16,270 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742198_1374 (size=71) 2025-01-08T05:01:16,270 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742198_1374 (size=71) 2025-01-08T05:01:16,270 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742198_1374 (size=71) 2025-01-08T05:01:16,271 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=143}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed. 2025-01-08T05:01:16,271 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=143}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=143 2025-01-08T05:01:16,272 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=143 2025-01-08T05:01:16,272 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportExpiredSnapshot on region 8ddd1acb9e69eae44cba4b9604557fed 2025-01-08T05:01:16,272 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=143, ppid=141, state=RUNNABLE; SnapshotRegionProcedure 8ddd1acb9e69eae44cba4b9604557fed 2025-01-08T05:01:16,275 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=143, resume processing ppid=141 2025-01-08T05:01:16,275 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=143, ppid=141, state=SUCCESS; SnapshotRegionProcedure 8ddd1acb9e69eae44cba4b9604557fed in 180 msec 2025-01-08T05:01:16,275 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T05:01:16,276 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T05:01:16,277 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T05:01:16,277 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T05:01:16,277 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:01:16,277 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(328): No files under family: cf 2025-01-08T05:01:16,291 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742199_1375 (size=63) 2025-01-08T05:01:16,292 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742199_1375 (size=63) 2025-01-08T05:01:16,292 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742199_1375 (size=63) 2025-01-08T05:01:16,295 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T05:01:16,295 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for emptySnaptb0-testExportExpiredSnapshot 2025-01-08T05:01:16,296 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/emptySnaptb0-testExportExpiredSnapshot 2025-01-08T05:01:16,318 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742200_1376 (size=653) 2025-01-08T05:01:16,319 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742200_1376 (size=653) 2025-01-08T05:01:16,319 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742200_1376 (size=653) 2025-01-08T05:01:16,323 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T05:01:16,327 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T05:01:16,328 DEBUG [PEWorker-1 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/emptySnaptb0-testExportExpiredSnapshot to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/emptySnaptb0-testExportExpiredSnapshot 2025-01-08T05:01:16,329 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T05:01:16,329 DEBUG [PEWorker-1 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 }, snapshot procedure id = 141 2025-01-08T05:01:16,330 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=141, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } in 260 msec 2025-01-08T05:01:16,373 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=141 2025-01-08T05:01:16,373 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportExpiredSnapshot, procId: 141 completed 2025-01-08T05:01:16,377 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37967 {}] regionserver.HRegion(8254): writing data to region testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T05:01:16,378 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39801 {}] regionserver.HRegion(8254): writing data to region testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T05:01:16,383 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportExpiredSnapshot 2025-01-08T05:01:16,383 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec. 2025-01-08T05:01:16,383 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T05:01:16,400 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } 2025-01-08T05:01:16,400 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736312476400 (current time:1736312476400). 2025-01-08T05:01:16,400 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T05:01:16,400 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testExportExpiredSnapshot VERSION not specified, setting to 2 2025-01-08T05:01:16,400 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T05:01:16,401 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x6add103c to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@f9b9245 2025-01-08T05:01:16,409 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1daaa168, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:01:16,411 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:01:16,412 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:43192, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:01:16,413 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x6add103c to 127.0.0.1:59048 2025-01-08T05:01:16,413 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:01:16,415 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x59c569da to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@a9fd11d 2025-01-08T05:01:16,423 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@42f690c7, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:01:16,427 DEBUG [hconnection-0x479b861a-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:01:16,428 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:43194, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:01:16,430 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x59c569da to 127.0.0.1:59048 2025-01-08T05:01:16,430 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:01:16,430 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportExpiredSnapshot], kv [jenkins: RWXCA] 2025-01-08T05:01:16,431 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T05:01:16,432 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=144, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } 2025-01-08T05:01:16,432 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 }, snapshot procedure id = 144 2025-01-08T05:01:16,433 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=144 2025-01-08T05:01:16,433 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T05:01:16,434 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T05:01:16,436 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T05:01:16,464 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742201_1377 (size=165) 2025-01-08T05:01:16,464 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742201_1377 (size=165) 2025-01-08T05:01:16,464 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742201_1377 (size=165) 2025-01-08T05:01:16,468 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T05:01:16,468 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=145, ppid=144, state=RUNNABLE; SnapshotRegionProcedure 4718f4a49bc4b9e61ea5fdb2dbd018ec}, {pid=146, ppid=144, state=RUNNABLE; SnapshotRegionProcedure 8ddd1acb9e69eae44cba4b9604557fed}] 2025-01-08T05:01:16,471 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=146, ppid=144, state=RUNNABLE; SnapshotRegionProcedure 8ddd1acb9e69eae44cba4b9604557fed 2025-01-08T05:01:16,472 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=145, ppid=144, state=RUNNABLE; SnapshotRegionProcedure 4718f4a49bc4b9e61ea5fdb2dbd018ec 2025-01-08T05:01:16,534 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=144 2025-01-08T05:01:16,624 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,39801,1736312288779 2025-01-08T05:01:16,624 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=39801 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=146 2025-01-08T05:01:16,625 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed. 2025-01-08T05:01:16,625 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:01:16,625 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.HRegion(2837): Flushing 8ddd1acb9e69eae44cba4b9604557fed 1/1 column families, dataSize=3.00 KB heapSize=6.72 KB 2025-01-08T05:01:16,625 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37967 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=145 2025-01-08T05:01:16,625 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec. 2025-01-08T05:01:16,626 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.HRegion(2837): Flushing 4718f4a49bc4b9e61ea5fdb2dbd018ec 1/1 column families, dataSize=266 B heapSize=832 B 2025-01-08T05:01:16,649 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e202501089096e8f31a3c4c5d8181489393cd9868_4718f4a49bc4b9e61ea5fdb2dbd018ec is 71, key is 05aa2d008fa1e215c9b79ffef9b90b5a/cf:q/1736312476377/Put/seqid=0 2025-01-08T05:01:16,650 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b202501089c4fee3576834481990ae166c6554892_8ddd1acb9e69eae44cba4b9604557fed is 71, key is 19b13748f3a66d54f00a0ebd37ddbcc8/cf:q/1736312476378/Put/seqid=0 2025-01-08T05:01:16,656 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742202_1378 (size=5172) 2025-01-08T05:01:16,656 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742202_1378 (size=5172) 2025-01-08T05:01:16,657 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742202_1378 (size=5172) 2025-01-08T05:01:16,657 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:01:16,662 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e202501089096e8f31a3c4c5d8181489393cd9868_4718f4a49bc4b9e61ea5fdb2dbd018ec to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/d41d8cd98f00b204e9800998ecf8427e202501089096e8f31a3c4c5d8181489393cd9868_4718f4a49bc4b9e61ea5fdb2dbd018ec 2025-01-08T05:01:16,663 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/4718f4a49bc4b9e61ea5fdb2dbd018ec/.tmp/cf/b180c03985f0444fa51048860d8ddae3, store: [table=testtb-testExportExpiredSnapshot family=cf region=4718f4a49bc4b9e61ea5fdb2dbd018ec] 2025-01-08T05:01:16,664 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/4718f4a49bc4b9e61ea5fdb2dbd018ec/.tmp/cf/b180c03985f0444fa51048860d8ddae3 is 209, key is 0e50f155cbc8f28446e9b14c918014c6c/cf:q/1736312476377/Put/seqid=0 2025-01-08T05:01:16,666 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742203_1379 (size=8102) 2025-01-08T05:01:16,667 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742203_1379 (size=8102) 2025-01-08T05:01:16,667 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742203_1379 (size=8102) 2025-01-08T05:01:16,668 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:01:16,672 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b202501089c4fee3576834481990ae166c6554892_8ddd1acb9e69eae44cba4b9604557fed to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/c4ca4238a0b923820dcc509a6f75849b202501089c4fee3576834481990ae166c6554892_8ddd1acb9e69eae44cba4b9604557fed 2025-01-08T05:01:16,673 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/8ddd1acb9e69eae44cba4b9604557fed/.tmp/cf/aa1efe517df740fd86d01ab08c377245, store: [table=testtb-testExportExpiredSnapshot family=cf region=8ddd1acb9e69eae44cba4b9604557fed] 2025-01-08T05:01:16,674 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/8ddd1acb9e69eae44cba4b9604557fed/.tmp/cf/aa1efe517df740fd86d01ab08c377245 is 209, key is 191e1f883390d3c83ff2e7df9f4ddac1b/cf:q/1736312476378/Put/seqid=0 2025-01-08T05:01:16,681 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742204_1380 (size=6123) 2025-01-08T05:01:16,681 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742204_1380 (size=6123) 2025-01-08T05:01:16,681 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742204_1380 (size=6123) 2025-01-08T05:01:16,682 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=266, hasBloomFilter=true, into tmp file hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/4718f4a49bc4b9e61ea5fdb2dbd018ec/.tmp/cf/b180c03985f0444fa51048860d8ddae3 2025-01-08T05:01:16,688 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/4718f4a49bc4b9e61ea5fdb2dbd018ec/.tmp/cf/b180c03985f0444fa51048860d8ddae3 as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/4718f4a49bc4b9e61ea5fdb2dbd018ec/cf/b180c03985f0444fa51048860d8ddae3 2025-01-08T05:01:16,689 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742205_1381 (size=14794) 2025-01-08T05:01:16,689 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742205_1381 (size=14794) 2025-01-08T05:01:16,690 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742205_1381 (size=14794) 2025-01-08T05:01:16,690 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=3.0 K, hasBloomFilter=true, into tmp file hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/8ddd1acb9e69eae44cba4b9604557fed/.tmp/cf/aa1efe517df740fd86d01ab08c377245 2025-01-08T05:01:16,694 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/4718f4a49bc4b9e61ea5fdb2dbd018ec/cf/b180c03985f0444fa51048860d8ddae3, entries=4, sequenceid=6, filesize=6.0 K 2025-01-08T05:01:16,695 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.HRegion(3040): Finished flush of dataSize ~266 B/266, heapSize ~816 B/816, currentSize=0 B/0 for 4718f4a49bc4b9e61ea5fdb2dbd018ec in 69ms, sequenceid=6, compaction requested=false 2025-01-08T05:01:16,695 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportExpiredSnapshot' 2025-01-08T05:01:16,696 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/8ddd1acb9e69eae44cba4b9604557fed/.tmp/cf/aa1efe517df740fd86d01ab08c377245 as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/8ddd1acb9e69eae44cba4b9604557fed/cf/aa1efe517df740fd86d01ab08c377245 2025-01-08T05:01:16,696 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.HRegion(2538): Flush status journal for 4718f4a49bc4b9e61ea5fdb2dbd018ec: 2025-01-08T05:01:16,696 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec. for snaptb0-testExportExpiredSnapshot completed. 2025-01-08T05:01:16,696 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec.' region-info for snapshot=snaptb0-testExportExpiredSnapshot 2025-01-08T05:01:16,696 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T05:01:16,696 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/4718f4a49bc4b9e61ea5fdb2dbd018ec/cf/b180c03985f0444fa51048860d8ddae3] hfiles 2025-01-08T05:01:16,696 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/4718f4a49bc4b9e61ea5fdb2dbd018ec/cf/b180c03985f0444fa51048860d8ddae3 for snapshot=snaptb0-testExportExpiredSnapshot 2025-01-08T05:01:16,701 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/8ddd1acb9e69eae44cba4b9604557fed/cf/aa1efe517df740fd86d01ab08c377245, entries=46, sequenceid=6, filesize=14.4 K 2025-01-08T05:01:16,702 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.HRegion(3040): Finished flush of dataSize ~3.00 KB/3070, heapSize ~6.70 KB/6864, currentSize=0 B/0 for 8ddd1acb9e69eae44cba4b9604557fed in 77ms, sequenceid=6, compaction requested=false 2025-01-08T05:01:16,702 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.HRegion(2538): Flush status journal for 8ddd1acb9e69eae44cba4b9604557fed: 2025-01-08T05:01:16,702 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed. for snaptb0-testExportExpiredSnapshot completed. 2025-01-08T05:01:16,702 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed.' region-info for snapshot=snaptb0-testExportExpiredSnapshot 2025-01-08T05:01:16,702 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T05:01:16,702 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/8ddd1acb9e69eae44cba4b9604557fed/cf/aa1efe517df740fd86d01ab08c377245] hfiles 2025-01-08T05:01:16,702 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/8ddd1acb9e69eae44cba4b9604557fed/cf/aa1efe517df740fd86d01ab08c377245 for snapshot=snaptb0-testExportExpiredSnapshot 2025-01-08T05:01:16,706 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742206_1382 (size=110) 2025-01-08T05:01:16,706 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742206_1382 (size=110) 2025-01-08T05:01:16,707 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742206_1382 (size=110) 2025-01-08T05:01:16,707 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec. 2025-01-08T05:01:16,707 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=145 2025-01-08T05:01:16,707 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=145 2025-01-08T05:01:16,708 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportExpiredSnapshot on region 4718f4a49bc4b9e61ea5fdb2dbd018ec 2025-01-08T05:01:16,708 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=145, ppid=144, state=RUNNABLE; SnapshotRegionProcedure 4718f4a49bc4b9e61ea5fdb2dbd018ec 2025-01-08T05:01:16,709 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=145, ppid=144, state=SUCCESS; SnapshotRegionProcedure 4718f4a49bc4b9e61ea5fdb2dbd018ec in 240 msec 2025-01-08T05:01:16,720 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742207_1383 (size=110) 2025-01-08T05:01:16,720 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742207_1383 (size=110) 2025-01-08T05:01:16,720 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742207_1383 (size=110) 2025-01-08T05:01:16,721 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed. 2025-01-08T05:01:16,721 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=146 2025-01-08T05:01:16,722 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=146 2025-01-08T05:01:16,722 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportExpiredSnapshot on region 8ddd1acb9e69eae44cba4b9604557fed 2025-01-08T05:01:16,722 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=146, ppid=144, state=RUNNABLE; SnapshotRegionProcedure 8ddd1acb9e69eae44cba4b9604557fed 2025-01-08T05:01:16,725 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=146, resume processing ppid=144 2025-01-08T05:01:16,725 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=146, ppid=144, state=SUCCESS; SnapshotRegionProcedure 8ddd1acb9e69eae44cba4b9604557fed in 255 msec 2025-01-08T05:01:16,725 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T05:01:16,726 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T05:01:16,727 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T05:01:16,727 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T05:01:16,727 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:01:16,729 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/c4ca4238a0b923820dcc509a6f75849b202501089c4fee3576834481990ae166c6554892_8ddd1acb9e69eae44cba4b9604557fed, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/d41d8cd98f00b204e9800998ecf8427e202501089096e8f31a3c4c5d8181489393cd9868_4718f4a49bc4b9e61ea5fdb2dbd018ec] hfiles 2025-01-08T05:01:16,729 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/c4ca4238a0b923820dcc509a6f75849b202501089c4fee3576834481990ae166c6554892_8ddd1acb9e69eae44cba4b9604557fed 2025-01-08T05:01:16,729 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/d41d8cd98f00b204e9800998ecf8427e202501089096e8f31a3c4c5d8181489393cd9868_4718f4a49bc4b9e61ea5fdb2dbd018ec 2025-01-08T05:01:16,735 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=144 2025-01-08T05:01:16,744 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742208_1384 (size=294) 2025-01-08T05:01:16,744 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742208_1384 (size=294) 2025-01-08T05:01:16,745 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742208_1384 (size=294) 2025-01-08T05:01:16,746 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T05:01:16,746 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testExportExpiredSnapshot 2025-01-08T05:01:16,747 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/snaptb0-testExportExpiredSnapshot 2025-01-08T05:01:16,759 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742209_1385 (size=963) 2025-01-08T05:01:16,759 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742209_1385 (size=963) 2025-01-08T05:01:16,760 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742209_1385 (size=963) 2025-01-08T05:01:16,768 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T05:01:16,775 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T05:01:16,775 DEBUG [PEWorker-1 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/snaptb0-testExportExpiredSnapshot to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportExpiredSnapshot 2025-01-08T05:01:16,777 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T05:01:16,777 DEBUG [PEWorker-1 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 }, snapshot procedure id = 144 2025-01-08T05:01:16,779 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=144, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } in 346 msec 2025-01-08T05:01:17,036 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=144 2025-01-08T05:01:17,037 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportExpiredSnapshot, procId: 144 completed 2025-01-08T05:01:17,038 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.3 create 'testExportExpiredSnapshot', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T05:01:17,039 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=147, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testExportExpiredSnapshot 2025-01-08T05:01:17,041 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=147, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testExportExpiredSnapshot execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T05:01:17,041 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "testExportExpiredSnapshot" procId is: 147 2025-01-08T05:01:17,041 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=147 2025-01-08T05:01:17,042 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=147, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testExportExpiredSnapshot execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T05:01:17,061 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742210_1386 (size=436) 2025-01-08T05:01:17,061 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742210_1386 (size=436) 2025-01-08T05:01:17,061 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742210_1386 (size=436) 2025-01-08T05:01:17,064 INFO [RegionOpenAndInit-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => bf5538cd1c1443d97a8c76cd68b482f8, NAME => 'testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testExportExpiredSnapshot', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:01:17,064 INFO [RegionOpenAndInit-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => 5881ce671bd325587aadcdae5691632b, NAME => 'testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testExportExpiredSnapshot', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:01:17,075 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742211_1387 (size=61) 2025-01-08T05:01:17,075 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742211_1387 (size=61) 2025-01-08T05:01:17,076 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742211_1387 (size=61) 2025-01-08T05:01:17,076 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(894): Instantiated testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:01:17,076 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1681): Closing 5881ce671bd325587aadcdae5691632b, disabling compactions & flushes 2025-01-08T05:01:17,076 INFO [RegionOpenAndInit-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1703): Closing region testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b. 2025-01-08T05:01:17,077 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b. 2025-01-08T05:01:17,077 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b. after waiting 0 ms 2025-01-08T05:01:17,077 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b. 2025-01-08T05:01:17,077 INFO [RegionOpenAndInit-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1922): Closed testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b. 2025-01-08T05:01:17,077 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1635): Region close journal for 5881ce671bd325587aadcdae5691632b: 2025-01-08T05:01:17,077 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742212_1388 (size=61) 2025-01-08T05:01:17,077 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742212_1388 (size=61) 2025-01-08T05:01:17,078 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742212_1388 (size=61) 2025-01-08T05:01:17,090 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(894): Instantiated testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:01:17,090 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1681): Closing bf5538cd1c1443d97a8c76cd68b482f8, disabling compactions & flushes 2025-01-08T05:01:17,090 INFO [RegionOpenAndInit-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1703): Closing region testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8. 2025-01-08T05:01:17,090 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8. 2025-01-08T05:01:17,090 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8. after waiting 0 ms 2025-01-08T05:01:17,090 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8. 2025-01-08T05:01:17,090 INFO [RegionOpenAndInit-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1922): Closed testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8. 2025-01-08T05:01:17,090 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1635): Region close journal for bf5538cd1c1443d97a8c76cd68b482f8: 2025-01-08T05:01:17,096 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=147, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testExportExpiredSnapshot execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T05:01:17,096 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b.","families":{"info":[{"qualifier":"regioninfo","vlen":60,"tag":[],"timestamp":"1736312477096"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736312477096"}]},"ts":"1736312477096"} 2025-01-08T05:01:17,096 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8.","families":{"info":[{"qualifier":"regioninfo","vlen":60,"tag":[],"timestamp":"1736312477096"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736312477096"}]},"ts":"1736312477096"} 2025-01-08T05:01:17,098 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T05:01:17,099 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=147, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testExportExpiredSnapshot execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T05:01:17,100 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testExportExpiredSnapshot","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312477099"}]},"ts":"1736312477099"} 2025-01-08T05:01:17,103 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testExportExpiredSnapshot, state=ENABLING in hbase:meta 2025-01-08T05:01:17,107 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(202): Hosts are {f28094b72178=0} racks are {/default-rack=0} 2025-01-08T05:01:17,108 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T05:01:17,108 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T05:01:17,108 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T05:01:17,109 INFO [PEWorker-3 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T05:01:17,109 INFO [PEWorker-3 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T05:01:17,109 INFO [PEWorker-3 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T05:01:17,109 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T05:01:17,109 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=148, ppid=147, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testExportExpiredSnapshot, region=bf5538cd1c1443d97a8c76cd68b482f8, ASSIGN}, {pid=149, ppid=147, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testExportExpiredSnapshot, region=5881ce671bd325587aadcdae5691632b, ASSIGN}] 2025-01-08T05:01:17,110 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=149, ppid=147, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testExportExpiredSnapshot, region=5881ce671bd325587aadcdae5691632b, ASSIGN 2025-01-08T05:01:17,110 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=148, ppid=147, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testExportExpiredSnapshot, region=bf5538cd1c1443d97a8c76cd68b482f8, ASSIGN 2025-01-08T05:01:17,111 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=149, ppid=147, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testExportExpiredSnapshot, region=5881ce671bd325587aadcdae5691632b, ASSIGN; state=OFFLINE, location=f28094b72178,39801,1736312288779; forceNewPlan=false, retain=false 2025-01-08T05:01:17,111 INFO [PEWorker-5 {}] assignment.TransitRegionStateProcedure(264): Starting pid=148, ppid=147, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testExportExpiredSnapshot, region=bf5538cd1c1443d97a8c76cd68b482f8, ASSIGN; state=OFFLINE, location=f28094b72178,37967,1736312288941; forceNewPlan=false, retain=false 2025-01-08T05:01:17,142 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=147 2025-01-08T05:01:17,261 INFO [f28094b72178:42195 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T05:01:17,261 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=149 updating hbase:meta row=5881ce671bd325587aadcdae5691632b, regionState=OPENING, regionLocation=f28094b72178,39801,1736312288779 2025-01-08T05:01:17,261 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=148 updating hbase:meta row=bf5538cd1c1443d97a8c76cd68b482f8, regionState=OPENING, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T05:01:17,263 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=150, ppid=149, state=RUNNABLE; OpenRegionProcedure 5881ce671bd325587aadcdae5691632b, server=f28094b72178,39801,1736312288779}] 2025-01-08T05:01:17,265 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=151, ppid=148, state=RUNNABLE; OpenRegionProcedure bf5538cd1c1443d97a8c76cd68b482f8, server=f28094b72178,37967,1736312288941}] 2025-01-08T05:01:17,343 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=147 2025-01-08T05:01:17,417 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,39801,1736312288779 2025-01-08T05:01:17,418 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:01:17,421 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] handler.AssignRegionHandler(135): Open testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b. 2025-01-08T05:01:17,421 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(7285): Opening region: {ENCODED => 5881ce671bd325587aadcdae5691632b, NAME => 'testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T05:01:17,421 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(7999): Registered coprocessor service: region=testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b. service=AccessControlService 2025-01-08T05:01:17,421 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] handler.AssignRegionHandler(135): Open testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8. 2025-01-08T05:01:17,421 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T05:01:17,422 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(7285): Opening region: {ENCODED => bf5538cd1c1443d97a8c76cd68b482f8, NAME => 'testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T05:01:17,422 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testExportExpiredSnapshot 5881ce671bd325587aadcdae5691632b 2025-01-08T05:01:17,422 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(894): Instantiated testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:01:17,422 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(7999): Registered coprocessor service: region=testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8. service=AccessControlService 2025-01-08T05:01:17,422 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(7327): checking encryption for 5881ce671bd325587aadcdae5691632b 2025-01-08T05:01:17,422 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(7330): checking classloading for 5881ce671bd325587aadcdae5691632b 2025-01-08T05:01:17,422 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T05:01:17,422 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testExportExpiredSnapshot bf5538cd1c1443d97a8c76cd68b482f8 2025-01-08T05:01:17,422 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(894): Instantiated testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:01:17,422 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(7327): checking encryption for bf5538cd1c1443d97a8c76cd68b482f8 2025-01-08T05:01:17,422 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(7330): checking classloading for bf5538cd1c1443d97a8c76cd68b482f8 2025-01-08T05:01:17,443 INFO [StoreOpener-bf5538cd1c1443d97a8c76cd68b482f8-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region bf5538cd1c1443d97a8c76cd68b482f8 2025-01-08T05:01:17,446 INFO [StoreOpener-bf5538cd1c1443d97a8c76cd68b482f8-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region bf5538cd1c1443d97a8c76cd68b482f8 columnFamilyName cf 2025-01-08T05:01:17,446 INFO [StoreOpener-5881ce671bd325587aadcdae5691632b-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 5881ce671bd325587aadcdae5691632b 2025-01-08T05:01:17,447 DEBUG [StoreOpener-bf5538cd1c1443d97a8c76cd68b482f8-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:01:17,448 INFO [StoreOpener-bf5538cd1c1443d97a8c76cd68b482f8-1 {}] regionserver.HStore(327): Store=bf5538cd1c1443d97a8c76cd68b482f8/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T05:01:17,449 INFO [StoreOpener-5881ce671bd325587aadcdae5691632b-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 5881ce671bd325587aadcdae5691632b columnFamilyName cf 2025-01-08T05:01:17,453 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportExpiredSnapshot/bf5538cd1c1443d97a8c76cd68b482f8 2025-01-08T05:01:17,454 DEBUG [StoreOpener-5881ce671bd325587aadcdae5691632b-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:01:17,454 INFO [StoreOpener-5881ce671bd325587aadcdae5691632b-1 {}] regionserver.HStore(327): Store=5881ce671bd325587aadcdae5691632b/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T05:01:17,455 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportExpiredSnapshot/bf5538cd1c1443d97a8c76cd68b482f8 2025-01-08T05:01:17,456 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportExpiredSnapshot/5881ce671bd325587aadcdae5691632b 2025-01-08T05:01:17,457 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportExpiredSnapshot/5881ce671bd325587aadcdae5691632b 2025-01-08T05:01:17,459 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(1085): writing seq id for bf5538cd1c1443d97a8c76cd68b482f8 2025-01-08T05:01:17,459 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(1085): writing seq id for 5881ce671bd325587aadcdae5691632b 2025-01-08T05:01:17,467 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportExpiredSnapshot/bf5538cd1c1443d97a8c76cd68b482f8/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T05:01:17,467 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(1102): Opened bf5538cd1c1443d97a8c76cd68b482f8; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=69105698, jitterRate=0.02975514531135559}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T05:01:17,468 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(1001): Region open journal for bf5538cd1c1443d97a8c76cd68b482f8: 2025-01-08T05:01:17,469 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegionServer(2601): Post open deploy tasks for testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8., pid=151, masterSystemTime=1736312477418 2025-01-08T05:01:17,471 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportExpiredSnapshot/5881ce671bd325587aadcdae5691632b/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T05:01:17,472 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(1102): Opened 5881ce671bd325587aadcdae5691632b; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=73704372, jitterRate=0.09828072786331177}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T05:01:17,472 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegionServer(2628): Finished post open deploy task for testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8. 2025-01-08T05:01:17,472 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] handler.AssignRegionHandler(164): Opened testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8. 2025-01-08T05:01:17,472 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(1001): Region open journal for 5881ce671bd325587aadcdae5691632b: 2025-01-08T05:01:17,472 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=148 updating hbase:meta row=bf5538cd1c1443d97a8c76cd68b482f8, regionState=OPEN, openSeqNum=2, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T05:01:17,473 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegionServer(2601): Post open deploy tasks for testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b., pid=150, masterSystemTime=1736312477416 2025-01-08T05:01:17,475 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegionServer(2628): Finished post open deploy task for testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b. 2025-01-08T05:01:17,475 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] handler.AssignRegionHandler(164): Opened testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b. 2025-01-08T05:01:17,477 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=149 updating hbase:meta row=5881ce671bd325587aadcdae5691632b, regionState=OPEN, openSeqNum=2, regionLocation=f28094b72178,39801,1736312288779 2025-01-08T05:01:17,478 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=151, resume processing ppid=148 2025-01-08T05:01:17,478 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=151, ppid=148, state=SUCCESS; OpenRegionProcedure bf5538cd1c1443d97a8c76cd68b482f8, server=f28094b72178,37967,1736312288941 in 209 msec 2025-01-08T05:01:17,482 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=148, ppid=147, state=SUCCESS; TransitRegionStateProcedure table=testExportExpiredSnapshot, region=bf5538cd1c1443d97a8c76cd68b482f8, ASSIGN in 370 msec 2025-01-08T05:01:17,483 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=150, resume processing ppid=149 2025-01-08T05:01:17,484 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=150, ppid=149, state=SUCCESS; OpenRegionProcedure 5881ce671bd325587aadcdae5691632b, server=f28094b72178,39801,1736312288779 in 217 msec 2025-01-08T05:01:17,487 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=149, resume processing ppid=147 2025-01-08T05:01:17,487 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=149, ppid=147, state=SUCCESS; TransitRegionStateProcedure table=testExportExpiredSnapshot, region=5881ce671bd325587aadcdae5691632b, ASSIGN in 375 msec 2025-01-08T05:01:17,488 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=147, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testExportExpiredSnapshot execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T05:01:17,488 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testExportExpiredSnapshot","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312477488"}]},"ts":"1736312477488"} 2025-01-08T05:01:17,492 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testExportExpiredSnapshot, state=ENABLED in hbase:meta 2025-01-08T05:01:17,495 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=147, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testExportExpiredSnapshot execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T05:01:17,495 DEBUG [PEWorker-3 {}] access.PermissionStorage(175): Writing permission with rowKey testExportExpiredSnapshot jenkins: RWXCA 2025-01-08T05:01:17,498 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37583 {}] access.PermissionStorage(611): Read acl: entry[testExportExpiredSnapshot], kv [jenkins: RWXCA] 2025-01-08T05:01:17,501 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:17,501 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:17,501 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:17,501 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:17,505 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:17,505 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:17,505 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:17,506 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:17,506 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:17,506 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:17,506 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:17,507 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:17,512 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=147, state=SUCCESS; CreateTableProcedure table=testExportExpiredSnapshot in 466 msec 2025-01-08T05:01:17,645 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=147 2025-01-08T05:01:17,645 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testExportExpiredSnapshot, procId: 147 completed 2025-01-08T05:01:17,648 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testExportExpiredSnapshot 2025-01-08T05:01:17,648 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8. 2025-01-08T05:01:17,648 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T05:01:17,668 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37967 {}] regionserver.HRegion(8254): writing data to region testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T05:01:17,668 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39801 {}] regionserver.HRegion(8254): writing data to region testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T05:01:17,672 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testExportExpiredSnapshot 2025-01-08T05:01:17,672 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8. 2025-01-08T05:01:17,672 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T05:01:17,683 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } 2025-01-08T05:01:17,683 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snapshot-testExportExpiredSnapshot VERSION not specified, setting to 2 2025-01-08T05:01:17,683 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T05:01:17,684 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x0c11d011 to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@41fd7415 2025-01-08T05:01:17,688 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4c55f5cb, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:01:17,689 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:01:17,690 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:43208, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:01:17,691 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x0c11d011 to 127.0.0.1:59048 2025-01-08T05:01:17,691 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:01:17,692 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x5f9fd1ef to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@75e85bd5 2025-01-08T05:01:17,701 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@16f0030e, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:01:17,704 DEBUG [hconnection-0x2bd93e7c-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:01:17,706 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:43218, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:01:17,709 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x5f9fd1ef to 127.0.0.1:59048 2025-01-08T05:01:17,709 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:01:17,709 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] access.PermissionStorage(611): Read acl: entry[testExportExpiredSnapshot], kv [jenkins: RWXCA] 2025-01-08T05:01:17,710 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T05:01:17,711 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=152, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } 2025-01-08T05:01:17,711 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 }, snapshot procedure id = 152 2025-01-08T05:01:17,712 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_PREPARE 2025-01-08T05:01:17,712 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=152 2025-01-08T05:01:17,713 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T05:01:17,717 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T05:01:17,743 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742213_1389 (size=152) 2025-01-08T05:01:17,743 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742213_1389 (size=152) 2025-01-08T05:01:17,744 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742213_1389 (size=152) 2025-01-08T05:01:17,747 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T05:01:17,747 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=153, ppid=152, state=RUNNABLE; SnapshotRegionProcedure bf5538cd1c1443d97a8c76cd68b482f8}, {pid=154, ppid=152, state=RUNNABLE; SnapshotRegionProcedure 5881ce671bd325587aadcdae5691632b}] 2025-01-08T05:01:17,748 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=153, ppid=152, state=RUNNABLE; SnapshotRegionProcedure bf5538cd1c1443d97a8c76cd68b482f8 2025-01-08T05:01:17,748 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=154, ppid=152, state=RUNNABLE; SnapshotRegionProcedure 5881ce671bd325587aadcdae5691632b 2025-01-08T05:01:17,813 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=152 2025-01-08T05:01:17,899 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:01:17,900 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37967 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=153 2025-01-08T05:01:17,900 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,39801,1736312288779 2025-01-08T05:01:17,901 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=39801 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=154 2025-01-08T05:01:17,901 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b. 2025-01-08T05:01:17,902 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.HRegion(2837): Flushing 5881ce671bd325587aadcdae5691632b 1/1 column families, dataSize=2.61 KB heapSize=5.88 KB 2025-01-08T05:01:17,903 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8. 2025-01-08T05:01:17,903 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.HRegion(2837): Flushing bf5538cd1c1443d97a8c76cd68b482f8 1/1 column families, dataSize=668 B heapSize=1.66 KB 2025-01-08T05:01:17,917 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testExportExpiredSnapshot' 2025-01-08T05:01:17,953 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b202501082b8eea93269f46989af193fcf6d0b51c_5881ce671bd325587aadcdae5691632b is 71, key is 12561255db1ed0ebb96d5e96ae3a871a/cf:q/1736312477668/Put/seqid=0 2025-01-08T05:01:17,965 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e2025010805510957bddc4a3ebca114d1bbd703ef_bf5538cd1c1443d97a8c76cd68b482f8 is 71, key is 02960997948f9f71e2387e4ef4889cd9/cf:q/1736312477668/Put/seqid=0 2025-01-08T05:01:17,988 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742214_1390 (size=7681) 2025-01-08T05:01:17,988 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742214_1390 (size=7681) 2025-01-08T05:01:17,989 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742214_1390 (size=7681) 2025-01-08T05:01:17,990 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:01:17,996 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b202501082b8eea93269f46989af193fcf6d0b51c_5881ce671bd325587aadcdae5691632b to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testExportExpiredSnapshot/1030f41967fbb659ab4c2a7a1774d313/cf/c4ca4238a0b923820dcc509a6f75849b202501082b8eea93269f46989af193fcf6d0b51c_5881ce671bd325587aadcdae5691632b 2025-01-08T05:01:17,998 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportExpiredSnapshot/5881ce671bd325587aadcdae5691632b/.tmp/cf/77be662708464d09a389301e742b9e00, store: [table=testExportExpiredSnapshot family=cf region=5881ce671bd325587aadcdae5691632b] 2025-01-08T05:01:17,999 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportExpiredSnapshot/5881ce671bd325587aadcdae5691632b/.tmp/cf/77be662708464d09a389301e742b9e00 is 202, key is 189b46a0f45d76936c9c9fba13544b210/cf:q/1736312477668/Put/seqid=0 2025-01-08T05:01:18,004 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742215_1391 (size=5591) 2025-01-08T05:01:18,005 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742215_1391 (size=5591) 2025-01-08T05:01:18,006 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742215_1391 (size=5591) 2025-01-08T05:01:18,007 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:01:18,015 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e2025010805510957bddc4a3ebca114d1bbd703ef_bf5538cd1c1443d97a8c76cd68b482f8 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testExportExpiredSnapshot/1030f41967fbb659ab4c2a7a1774d313/cf/d41d8cd98f00b204e9800998ecf8427e2025010805510957bddc4a3ebca114d1bbd703ef_bf5538cd1c1443d97a8c76cd68b482f8 2025-01-08T05:01:18,015 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=152 2025-01-08T05:01:18,016 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportExpiredSnapshot/bf5538cd1c1443d97a8c76cd68b482f8/.tmp/cf/737886aa5ea44e05aad6c0d1fe7c12d2, store: [table=testExportExpiredSnapshot family=cf region=bf5538cd1c1443d97a8c76cd68b482f8] 2025-01-08T05:01:18,017 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportExpiredSnapshot/bf5538cd1c1443d97a8c76cd68b482f8/.tmp/cf/737886aa5ea44e05aad6c0d1fe7c12d2 is 202, key is 03b6028cedb9d284ce90f2dccdc4a91a8/cf:q/1736312477668/Put/seqid=0 2025-01-08T05:01:18,035 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742216_1392 (size=13275) 2025-01-08T05:01:18,035 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742216_1392 (size=13275) 2025-01-08T05:01:18,036 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742216_1392 (size=13275) 2025-01-08T05:01:18,037 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=5, memsize=2.6 K, hasBloomFilter=true, into tmp file hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportExpiredSnapshot/5881ce671bd325587aadcdae5691632b/.tmp/cf/77be662708464d09a389301e742b9e00 2025-01-08T05:01:18,044 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportExpiredSnapshot/5881ce671bd325587aadcdae5691632b/.tmp/cf/77be662708464d09a389301e742b9e00 as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportExpiredSnapshot/5881ce671bd325587aadcdae5691632b/cf/77be662708464d09a389301e742b9e00 2025-01-08T05:01:18,050 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportExpiredSnapshot/5881ce671bd325587aadcdae5691632b/cf/77be662708464d09a389301e742b9e00, entries=40, sequenceid=5, filesize=13.0 K 2025-01-08T05:01:18,051 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.HRegion(3040): Finished flush of dataSize ~2.61 KB/2668, heapSize ~5.86 KB/6000, currentSize=0 B/0 for 5881ce671bd325587aadcdae5691632b in 150ms, sequenceid=5, compaction requested=false 2025-01-08T05:01:18,051 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.HRegion(2538): Flush status journal for 5881ce671bd325587aadcdae5691632b: 2025-01-08T05:01:18,051 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.SnapshotRegionCallable(75): Snapshotting region testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b. for snapshot-testExportExpiredSnapshot completed. 2025-01-08T05:01:18,052 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] snapshot.SnapshotManifest(241): Storing 'testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b.' region-info for snapshot=snapshot-testExportExpiredSnapshot 2025-01-08T05:01:18,052 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T05:01:18,052 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportExpiredSnapshot/5881ce671bd325587aadcdae5691632b/cf/77be662708464d09a389301e742b9e00] hfiles 2025-01-08T05:01:18,052 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportExpiredSnapshot/5881ce671bd325587aadcdae5691632b/cf/77be662708464d09a389301e742b9e00 for snapshot=snapshot-testExportExpiredSnapshot 2025-01-08T05:01:18,065 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742217_1393 (size=7282) 2025-01-08T05:01:18,065 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742217_1393 (size=7282) 2025-01-08T05:01:18,066 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742217_1393 (size=7282) 2025-01-08T05:01:18,067 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=5, memsize=668, hasBloomFilter=true, into tmp file hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportExpiredSnapshot/bf5538cd1c1443d97a8c76cd68b482f8/.tmp/cf/737886aa5ea44e05aad6c0d1fe7c12d2 2025-01-08T05:01:18,074 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportExpiredSnapshot/bf5538cd1c1443d97a8c76cd68b482f8/.tmp/cf/737886aa5ea44e05aad6c0d1fe7c12d2 as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportExpiredSnapshot/bf5538cd1c1443d97a8c76cd68b482f8/cf/737886aa5ea44e05aad6c0d1fe7c12d2 2025-01-08T05:01:18,085 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportExpiredSnapshot/bf5538cd1c1443d97a8c76cd68b482f8/cf/737886aa5ea44e05aad6c0d1fe7c12d2, entries=10, sequenceid=5, filesize=7.1 K 2025-01-08T05:01:18,087 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.HRegion(3040): Finished flush of dataSize ~668 B/668, heapSize ~1.64 KB/1680, currentSize=0 B/0 for bf5538cd1c1443d97a8c76cd68b482f8 in 184ms, sequenceid=5, compaction requested=false 2025-01-08T05:01:18,087 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.HRegion(2538): Flush status journal for bf5538cd1c1443d97a8c76cd68b482f8: 2025-01-08T05:01:18,087 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.SnapshotRegionCallable(75): Snapshotting region testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8. for snapshot-testExportExpiredSnapshot completed. 2025-01-08T05:01:18,087 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] snapshot.SnapshotManifest(241): Storing 'testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8.' region-info for snapshot=snapshot-testExportExpiredSnapshot 2025-01-08T05:01:18,087 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T05:01:18,087 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportExpiredSnapshot/bf5538cd1c1443d97a8c76cd68b482f8/cf/737886aa5ea44e05aad6c0d1fe7c12d2] hfiles 2025-01-08T05:01:18,087 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportExpiredSnapshot/bf5538cd1c1443d97a8c76cd68b482f8/cf/737886aa5ea44e05aad6c0d1fe7c12d2 for snapshot=snapshot-testExportExpiredSnapshot 2025-01-08T05:01:18,132 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742218_1394 (size=103) 2025-01-08T05:01:18,132 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742218_1394 (size=103) 2025-01-08T05:01:18,133 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742218_1394 (size=103) 2025-01-08T05:01:18,133 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b. 2025-01-08T05:01:18,134 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=154 2025-01-08T05:01:18,134 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=154 2025-01-08T05:01:18,134 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snapshot-testExportExpiredSnapshot on region 5881ce671bd325587aadcdae5691632b 2025-01-08T05:01:18,134 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=154, ppid=152, state=RUNNABLE; SnapshotRegionProcedure 5881ce671bd325587aadcdae5691632b 2025-01-08T05:01:18,137 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=154, ppid=152, state=SUCCESS; SnapshotRegionProcedure 5881ce671bd325587aadcdae5691632b in 389 msec 2025-01-08T05:01:18,170 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742219_1395 (size=103) 2025-01-08T05:01:18,173 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8. 2025-01-08T05:01:18,173 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=153 2025-01-08T05:01:18,173 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=153 2025-01-08T05:01:18,173 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snapshot-testExportExpiredSnapshot on region bf5538cd1c1443d97a8c76cd68b482f8 2025-01-08T05:01:18,174 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=153, ppid=152, state=RUNNABLE; SnapshotRegionProcedure bf5538cd1c1443d97a8c76cd68b482f8 2025-01-08T05:01:18,174 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742219_1395 (size=103) 2025-01-08T05:01:18,174 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742219_1395 (size=103) 2025-01-08T05:01:18,179 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=153, resume processing ppid=152 2025-01-08T05:01:18,179 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=153, ppid=152, state=SUCCESS; SnapshotRegionProcedure bf5538cd1c1443d97a8c76cd68b482f8 in 429 msec 2025-01-08T05:01:18,179 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T05:01:18,180 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T05:01:18,182 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T05:01:18,182 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T05:01:18,182 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:01:18,183 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testExportExpiredSnapshot/1030f41967fbb659ab4c2a7a1774d313/cf/c4ca4238a0b923820dcc509a6f75849b202501082b8eea93269f46989af193fcf6d0b51c_5881ce671bd325587aadcdae5691632b, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testExportExpiredSnapshot/1030f41967fbb659ab4c2a7a1774d313/cf/d41d8cd98f00b204e9800998ecf8427e2025010805510957bddc4a3ebca114d1bbd703ef_bf5538cd1c1443d97a8c76cd68b482f8] hfiles 2025-01-08T05:01:18,183 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testExportExpiredSnapshot/1030f41967fbb659ab4c2a7a1774d313/cf/c4ca4238a0b923820dcc509a6f75849b202501082b8eea93269f46989af193fcf6d0b51c_5881ce671bd325587aadcdae5691632b 2025-01-08T05:01:18,183 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testExportExpiredSnapshot/1030f41967fbb659ab4c2a7a1774d313/cf/d41d8cd98f00b204e9800998ecf8427e2025010805510957bddc4a3ebca114d1bbd703ef_bf5538cd1c1443d97a8c76cd68b482f8 2025-01-08T05:01:18,205 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742220_1396 (size=287) 2025-01-08T05:01:18,206 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742220_1396 (size=287) 2025-01-08T05:01:18,206 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742220_1396 (size=287) 2025-01-08T05:01:18,208 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T05:01:18,208 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snapshot-testExportExpiredSnapshot 2025-01-08T05:01:18,209 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/snapshot-testExportExpiredSnapshot 2025-01-08T05:01:18,236 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742221_1397 (size=935) 2025-01-08T05:01:18,236 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742221_1397 (size=935) 2025-01-08T05:01:18,237 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742221_1397 (size=935) 2025-01-08T05:01:18,244 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T05:01:18,270 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T05:01:18,271 DEBUG [PEWorker-3 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/snapshot-testExportExpiredSnapshot to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snapshot-testExportExpiredSnapshot 2025-01-08T05:01:18,283 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T05:01:18,283 DEBUG [PEWorker-3 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 }, snapshot procedure id = 152 2025-01-08T05:01:18,286 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=152, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } in 573 msec 2025-01-08T05:01:18,317 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=152 2025-01-08T05:01:18,317 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testExportExpiredSnapshot, procId: 152 completed 2025-01-08T05:01:18,345 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testExportExpiredSnapshot 2025-01-08T05:01:18,345 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testExportExpiredSnapshot Metrics about Tables on a single HBase RegionServer 2025-01-08T05:01:18,346 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportExpiredSnapshot 2025-01-08T05:01:18,346 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportExpiredSnapshot Metrics about Tables on a single HBase RegionServer 2025-01-08T05:01:18,346 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T05:01:18,346 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T05:01:19,110 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0006_000001 (auth:SIMPLE) from 127.0.0.1:60466 2025-01-08T05:01:19,138 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_2/usercache/jenkins/appcache/application_1736312298603_0006/container_1736312298603_0006_01_000001/launch_container.sh] 2025-01-08T05:01:19,138 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_2/usercache/jenkins/appcache/application_1736312298603_0006/container_1736312298603_0006_01_000001/container_tokens] 2025-01-08T05:01:19,138 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_2/usercache/jenkins/appcache/application_1736312298603_0006/container_1736312298603_0006_01_000001/sysfs] 2025-01-08T05:01:20,200 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T05:01:28,329 INFO [Time-limited test {}] snapshot.TestExportSnapshot(468): HDFS export destination path: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312488329 2025-01-08T05:01:28,329 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=hdfs://localhost:45939, tgtDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312488329, rawTgtDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312488329, srcFsUri=hdfs://localhost:45939, srcDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:01:28,388 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:45939, inputRoot=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:01:28,388 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1149249371_22, ugi=jenkins (auth:SIMPLE)]], outputRoot=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312488329, skipTmp=false, initialOutputSnapshotDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312488329/.hbase-snapshot/.tmp/snapshot-testExportExpiredSnapshot 2025-01-08T05:01:28,395 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T05:01:28,397 ERROR [Time-limited test {}] util.AbstractHBaseTool(153): Error running command-line tool org.apache.hadoop.hbase.snapshot.SnapshotTTLExpiredException: TTL for snapshot 'snapshot-testExportExpiredSnapshot' has already expired. at org.apache.hadoop.hbase.snapshot.ExportSnapshot.verifySnapshot(ExportSnapshot.java:948) ~[classes/:?] at org.apache.hadoop.hbase.snapshot.ExportSnapshot.doWork(ExportSnapshot.java:1093) ~[classes/:?] at org.apache.hadoop.hbase.util.AbstractHBaseTool.run(AbstractHBaseTool.java:151) ~[hbase-common-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:82) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hbase.snapshot.TestExportSnapshot.runExportSnapshot(TestExportSnapshot.java:523) ~[test-classes/:?] at org.apache.hadoop.hbase.snapshot.TestExportSnapshot.testExportExpiredSnapshot(TestExportSnapshot.java:315) ~[test-classes/:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) ~[junit-4.13.2.jar:4.13.2] at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) ~[junit-4.13.2.jar:4.13.2] at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2025-01-08T05:01:28,398 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testExportExpiredSnapshot 2025-01-08T05:01:28,398 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.3 disable testtb-testExportExpiredSnapshot 2025-01-08T05:01:28,400 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=155, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testExportExpiredSnapshot 2025-01-08T05:01:28,402 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportExpiredSnapshot","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312488402"}]},"ts":"1736312488402"} 2025-01-08T05:01:28,403 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=155 2025-01-08T05:01:28,407 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportExpiredSnapshot, state=DISABLING in hbase:meta 2025-01-08T05:01:28,409 INFO [PEWorker-4 {}] procedure.DisableTableProcedure(284): Set testtb-testExportExpiredSnapshot to state=DISABLING 2025-01-08T05:01:28,411 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=156, ppid=155, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testExportExpiredSnapshot}] 2025-01-08T05:01:28,413 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=157, ppid=156, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=4718f4a49bc4b9e61ea5fdb2dbd018ec, UNASSIGN}, {pid=158, ppid=156, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=8ddd1acb9e69eae44cba4b9604557fed, UNASSIGN}] 2025-01-08T05:01:28,414 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=158, ppid=156, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=8ddd1acb9e69eae44cba4b9604557fed, UNASSIGN 2025-01-08T05:01:28,414 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=157, ppid=156, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=4718f4a49bc4b9e61ea5fdb2dbd018ec, UNASSIGN 2025-01-08T05:01:28,415 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=157 updating hbase:meta row=4718f4a49bc4b9e61ea5fdb2dbd018ec, regionState=CLOSING, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T05:01:28,415 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=158 updating hbase:meta row=8ddd1acb9e69eae44cba4b9604557fed, regionState=CLOSING, regionLocation=f28094b72178,39801,1736312288779 2025-01-08T05:01:28,420 DEBUG [PEWorker-1 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T05:01:28,421 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=159, ppid=158, state=RUNNABLE; CloseRegionProcedure 8ddd1acb9e69eae44cba4b9604557fed, server=f28094b72178,39801,1736312288779}] 2025-01-08T05:01:28,421 DEBUG [PEWorker-2 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T05:01:28,421 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=160, ppid=157, state=RUNNABLE; CloseRegionProcedure 4718f4a49bc4b9e61ea5fdb2dbd018ec, server=f28094b72178,37967,1736312288941}] 2025-01-08T05:01:28,507 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=155 2025-01-08T05:01:28,573 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,39801,1736312288779 2025-01-08T05:01:28,573 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:01:28,575 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] handler.UnassignRegionHandler(124): Close 4718f4a49bc4b9e61ea5fdb2dbd018ec 2025-01-08T05:01:28,575 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T05:01:28,575 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] regionserver.HRegion(1681): Closing 4718f4a49bc4b9e61ea5fdb2dbd018ec, disabling compactions & flushes 2025-01-08T05:01:28,575 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] regionserver.HRegion(1703): Closing region testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec. 2025-01-08T05:01:28,575 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec. 2025-01-08T05:01:28,575 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec. after waiting 0 ms 2025-01-08T05:01:28,575 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec. 2025-01-08T05:01:28,578 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] handler.UnassignRegionHandler(124): Close 8ddd1acb9e69eae44cba4b9604557fed 2025-01-08T05:01:28,578 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T05:01:28,578 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] regionserver.HRegion(1681): Closing 8ddd1acb9e69eae44cba4b9604557fed, disabling compactions & flushes 2025-01-08T05:01:28,578 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] regionserver.HRegion(1703): Closing region testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed. 2025-01-08T05:01:28,578 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed. 2025-01-08T05:01:28,578 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed. after waiting 0 ms 2025-01-08T05:01:28,578 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed. 2025-01-08T05:01:28,635 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/4718f4a49bc4b9e61ea5fdb2dbd018ec/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T05:01:28,636 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:01:28,636 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] regionserver.HRegion(1922): Closed testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec. 2025-01-08T05:01:28,636 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] regionserver.HRegion(1635): Region close journal for 4718f4a49bc4b9e61ea5fdb2dbd018ec: 2025-01-08T05:01:28,647 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] handler.UnassignRegionHandler(170): Closed 4718f4a49bc4b9e61ea5fdb2dbd018ec 2025-01-08T05:01:28,648 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=157 updating hbase:meta row=4718f4a49bc4b9e61ea5fdb2dbd018ec, regionState=CLOSED 2025-01-08T05:01:28,653 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=160, resume processing ppid=157 2025-01-08T05:01:28,654 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=160, ppid=157, state=SUCCESS; CloseRegionProcedure 4718f4a49bc4b9e61ea5fdb2dbd018ec, server=f28094b72178,37967,1736312288941 in 229 msec 2025-01-08T05:01:28,654 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=157, ppid=156, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=4718f4a49bc4b9e61ea5fdb2dbd018ec, UNASSIGN in 240 msec 2025-01-08T05:01:28,664 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/8ddd1acb9e69eae44cba4b9604557fed/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T05:01:28,665 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:01:28,665 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] regionserver.HRegion(1922): Closed testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed. 2025-01-08T05:01:28,665 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] regionserver.HRegion(1635): Region close journal for 8ddd1acb9e69eae44cba4b9604557fed: 2025-01-08T05:01:28,668 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] handler.UnassignRegionHandler(170): Closed 8ddd1acb9e69eae44cba4b9604557fed 2025-01-08T05:01:28,669 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=158 updating hbase:meta row=8ddd1acb9e69eae44cba4b9604557fed, regionState=CLOSED 2025-01-08T05:01:28,673 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=159, resume processing ppid=158 2025-01-08T05:01:28,674 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=159, ppid=158, state=SUCCESS; CloseRegionProcedure 8ddd1acb9e69eae44cba4b9604557fed, server=f28094b72178,39801,1736312288779 in 251 msec 2025-01-08T05:01:28,679 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=158, resume processing ppid=156 2025-01-08T05:01:28,679 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=158, ppid=156, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=8ddd1acb9e69eae44cba4b9604557fed, UNASSIGN in 261 msec 2025-01-08T05:01:28,681 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=156, resume processing ppid=155 2025-01-08T05:01:28,681 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=156, ppid=155, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testExportExpiredSnapshot in 269 msec 2025-01-08T05:01:28,683 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportExpiredSnapshot","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312488682"}]},"ts":"1736312488682"} 2025-01-08T05:01:28,684 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportExpiredSnapshot, state=DISABLED in hbase:meta 2025-01-08T05:01:28,687 INFO [PEWorker-5 {}] procedure.DisableTableProcedure(296): Set testtb-testExportExpiredSnapshot to state=DISABLED 2025-01-08T05:01:28,693 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=155, state=SUCCESS; DisableTableProcedure table=testtb-testExportExpiredSnapshot in 291 msec 2025-01-08T05:01:28,708 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=155 2025-01-08T05:01:28,708 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testExportExpiredSnapshot, procId: 155 completed 2025-01-08T05:01:28,709 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.3 delete testtb-testExportExpiredSnapshot 2025-01-08T05:01:28,715 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=161, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testExportExpiredSnapshot 2025-01-08T05:01:28,716 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=161, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportExpiredSnapshot 2025-01-08T05:01:28,716 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testExportExpiredSnapshot 2025-01-08T05:01:28,718 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37583 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testExportExpiredSnapshot 2025-01-08T05:01:28,721 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=161, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testExportExpiredSnapshot 2025-01-08T05:01:28,722 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportExpiredSnapshot 2025-01-08T05:01:28,723 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportExpiredSnapshot 2025-01-08T05:01:28,723 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportExpiredSnapshot 2025-01-08T05:01:28,723 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportExpiredSnapshot 2025-01-08T05:01:28,724 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF 2025-01-08T05:01:28,724 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF 2025-01-08T05:01:28,724 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF 2025-01-08T05:01:28,724 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF 2025-01-08T05:01:28,727 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportExpiredSnapshot 2025-01-08T05:01:28,727 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportExpiredSnapshot 2025-01-08T05:01:28,727 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:28,727 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:28,728 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportExpiredSnapshot 2025-01-08T05:01:28,728 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:28,728 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportExpiredSnapshot 2025-01-08T05:01:28,728 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:28,728 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=161 2025-01-08T05:01:28,729 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:28,729 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:28,730 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:28,731 DEBUG [HFileArchiver-16 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/4718f4a49bc4b9e61ea5fdb2dbd018ec 2025-01-08T05:01:28,731 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:28,735 DEBUG [HFileArchiver-16 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/4718f4a49bc4b9e61ea5fdb2dbd018ec/cf, FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/4718f4a49bc4b9e61ea5fdb2dbd018ec/recovered.edits] 2025-01-08T05:01:28,741 DEBUG [HFileArchiver-17 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/8ddd1acb9e69eae44cba4b9604557fed 2025-01-08T05:01:28,748 DEBUG [HFileArchiver-17 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/8ddd1acb9e69eae44cba4b9604557fed/cf, FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/8ddd1acb9e69eae44cba4b9604557fed/recovered.edits] 2025-01-08T05:01:28,750 DEBUG [HFileArchiver-16 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/4718f4a49bc4b9e61ea5fdb2dbd018ec/cf/b180c03985f0444fa51048860d8ddae3 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportExpiredSnapshot/4718f4a49bc4b9e61ea5fdb2dbd018ec/cf/b180c03985f0444fa51048860d8ddae3 2025-01-08T05:01:28,754 DEBUG [HFileArchiver-17 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/8ddd1acb9e69eae44cba4b9604557fed/cf/aa1efe517df740fd86d01ab08c377245 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportExpiredSnapshot/8ddd1acb9e69eae44cba4b9604557fed/cf/aa1efe517df740fd86d01ab08c377245 2025-01-08T05:01:28,758 DEBUG [HFileArchiver-16 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/4718f4a49bc4b9e61ea5fdb2dbd018ec/recovered.edits/9.seqid to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportExpiredSnapshot/4718f4a49bc4b9e61ea5fdb2dbd018ec/recovered.edits/9.seqid 2025-01-08T05:01:28,760 DEBUG [HFileArchiver-16 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/4718f4a49bc4b9e61ea5fdb2dbd018ec 2025-01-08T05:01:28,760 DEBUG [HFileArchiver-17 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/8ddd1acb9e69eae44cba4b9604557fed/recovered.edits/9.seqid to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportExpiredSnapshot/8ddd1acb9e69eae44cba4b9604557fed/recovered.edits/9.seqid 2025-01-08T05:01:28,761 DEBUG [HFileArchiver-17 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportExpiredSnapshot/8ddd1acb9e69eae44cba4b9604557fed 2025-01-08T05:01:28,761 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(313): Archived testtb-testExportExpiredSnapshot regions 2025-01-08T05:01:28,762 DEBUG [PEWorker-3 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f 2025-01-08T05:01:28,763 DEBUG [PEWorker-3 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf] 2025-01-08T05:01:28,767 DEBUG [PEWorker-3 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/c4ca4238a0b923820dcc509a6f75849b202501089c4fee3576834481990ae166c6554892_8ddd1acb9e69eae44cba4b9604557fed to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/c4ca4238a0b923820dcc509a6f75849b202501089c4fee3576834481990ae166c6554892_8ddd1acb9e69eae44cba4b9604557fed 2025-01-08T05:01:28,768 DEBUG [PEWorker-3 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/d41d8cd98f00b204e9800998ecf8427e202501089096e8f31a3c4c5d8181489393cd9868_4718f4a49bc4b9e61ea5fdb2dbd018ec to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/d41d8cd98f00b204e9800998ecf8427e202501089096e8f31a3c4c5d8181489393cd9868_4718f4a49bc4b9e61ea5fdb2dbd018ec 2025-01-08T05:01:28,769 DEBUG [PEWorker-3 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f 2025-01-08T05:01:28,772 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=161, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testExportExpiredSnapshot 2025-01-08T05:01:28,790 WARN [PEWorker-3 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testtb-testExportExpiredSnapshot from hbase:meta 2025-01-08T05:01:28,793 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testExportExpiredSnapshot' descriptor. 2025-01-08T05:01:28,794 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=161, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testExportExpiredSnapshot 2025-01-08T05:01:28,794 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testExportExpiredSnapshot' from region states. 2025-01-08T05:01:28,794 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736312488794"}]},"ts":"9223372036854775807"} 2025-01-08T05:01:28,795 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736312488794"}]},"ts":"9223372036854775807"} 2025-01-08T05:01:28,809 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T05:01:28,809 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => 4718f4a49bc4b9e61ea5fdb2dbd018ec, NAME => 'testtb-testExportExpiredSnapshot,,1736312475418.4718f4a49bc4b9e61ea5fdb2dbd018ec.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => 8ddd1acb9e69eae44cba4b9604557fed, NAME => 'testtb-testExportExpiredSnapshot,1,1736312475418.8ddd1acb9e69eae44cba4b9604557fed.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T05:01:28,809 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testExportExpiredSnapshot' as deleted. 2025-01-08T05:01:28,809 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportExpiredSnapshot","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736312488809"}]},"ts":"9223372036854775807"} 2025-01-08T05:01:28,812 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testExportExpiredSnapshot state from META 2025-01-08T05:01:28,815 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(133): Finished pid=161, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportExpiredSnapshot 2025-01-08T05:01:28,817 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=161, state=SUCCESS; DeleteTableProcedure table=testtb-testExportExpiredSnapshot in 107 msec 2025-01-08T05:01:28,829 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=161 2025-01-08T05:01:28,830 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testExportExpiredSnapshot, procId: 161 completed 2025-01-08T05:01:28,849 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "emptySnaptb0-testExportExpiredSnapshot" 2025-01-08T05:01:28,852 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(380): Deleting snapshot: emptySnaptb0-testExportExpiredSnapshot 2025-01-08T05:01:28,853 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "snapshot-testExportExpiredSnapshot" 2025-01-08T05:01:28,855 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(380): Deleting snapshot: snapshot-testExportExpiredSnapshot 2025-01-08T05:01:28,857 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "snaptb0-testExportExpiredSnapshot" 2025-01-08T05:01:28,859 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testExportExpiredSnapshot 2025-01-08T05:01:28,889 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: snapshot.TestMobSecureExportSnapshot#testExportExpiredSnapshot Thread=796 (was 799), OpenFileDescriptor=797 (was 813), MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=882 (was 952), ProcessCount=16 (was 14) - ProcessCount LEAK? -, AvailableMemoryMB=2102 (was 2441) 2025-01-08T05:01:28,889 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=796 is superior to 500 2025-01-08T05:01:28,919 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: snapshot.TestMobSecureExportSnapshot#testEmptyExportFileSystemState Thread=796, OpenFileDescriptor=797, MaxFileDescriptor=1048576, SystemLoadAverage=882, ProcessCount=16, AvailableMemoryMB=2100 2025-01-08T05:01:28,919 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=796 is superior to 500 2025-01-08T05:01:28,921 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.3 create 'testtb-testEmptyExportFileSystemState', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T05:01:28,922 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=162, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testEmptyExportFileSystemState 2025-01-08T05:01:28,923 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=162, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testEmptyExportFileSystemState execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T05:01:28,923 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "testtb-testEmptyExportFileSystemState" procId is: 162 2025-01-08T05:01:28,924 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=162, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testEmptyExportFileSystemState execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T05:01:28,924 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=162 2025-01-08T05:01:28,967 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742222_1398 (size=448) 2025-01-08T05:01:28,967 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742222_1398 (size=448) 2025-01-08T05:01:28,967 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742222_1398 (size=448) 2025-01-08T05:01:28,971 INFO [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => e2b08e9e5645f931a0a5276a05929aba, NAME => 'testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testtb-testEmptyExportFileSystemState', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:01:28,972 INFO [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => f962e4f9abbc7826ddc0ddd4343efcb3, NAME => 'testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testtb-testEmptyExportFileSystemState', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:01:29,004 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742223_1399 (size=73) 2025-01-08T05:01:29,005 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742223_1399 (size=73) 2025-01-08T05:01:29,005 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742223_1399 (size=73) 2025-01-08T05:01:29,006 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742224_1400 (size=73) 2025-01-08T05:01:29,006 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742224_1400 (size=73) 2025-01-08T05:01:29,006 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:01:29,007 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-1 {}] regionserver.HRegion(1681): Closing f962e4f9abbc7826ddc0ddd4343efcb3, disabling compactions & flushes 2025-01-08T05:01:29,007 INFO [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3. 2025-01-08T05:01:29,007 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3. 2025-01-08T05:01:29,007 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3. after waiting 0 ms 2025-01-08T05:01:29,007 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3. 2025-01-08T05:01:29,007 INFO [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3. 2025-01-08T05:01:29,007 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742224_1400 (size=73) 2025-01-08T05:01:29,007 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-1 {}] regionserver.HRegion(1635): Region close journal for f962e4f9abbc7826ddc0ddd4343efcb3: 2025-01-08T05:01:29,008 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:01:29,008 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-0 {}] regionserver.HRegion(1681): Closing e2b08e9e5645f931a0a5276a05929aba, disabling compactions & flushes 2025-01-08T05:01:29,008 INFO [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba. 2025-01-08T05:01:29,008 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba. 2025-01-08T05:01:29,008 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba. after waiting 0 ms 2025-01-08T05:01:29,008 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba. 2025-01-08T05:01:29,008 INFO [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba. 2025-01-08T05:01:29,008 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-0 {}] regionserver.HRegion(1635): Region close journal for e2b08e9e5645f931a0a5276a05929aba: 2025-01-08T05:01:29,012 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=162, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testEmptyExportFileSystemState execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T05:01:29,012 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3.","families":{"info":[{"qualifier":"regioninfo","vlen":72,"tag":[],"timestamp":"1736312489012"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736312489012"}]},"ts":"1736312489012"} 2025-01-08T05:01:29,012 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba.","families":{"info":[{"qualifier":"regioninfo","vlen":72,"tag":[],"timestamp":"1736312489012"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736312489012"}]},"ts":"1736312489012"} 2025-01-08T05:01:29,015 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T05:01:29,017 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=162, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testEmptyExportFileSystemState execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T05:01:29,017 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testEmptyExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312489017"}]},"ts":"1736312489017"} 2025-01-08T05:01:29,018 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testEmptyExportFileSystemState, state=ENABLING in hbase:meta 2025-01-08T05:01:29,022 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(202): Hosts are {f28094b72178=0} racks are {/default-rack=0} 2025-01-08T05:01:29,024 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T05:01:29,024 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T05:01:29,024 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T05:01:29,024 INFO [PEWorker-2 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T05:01:29,024 INFO [PEWorker-2 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T05:01:29,024 INFO [PEWorker-2 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T05:01:29,024 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T05:01:29,024 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=163, ppid=162, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=e2b08e9e5645f931a0a5276a05929aba, ASSIGN}, {pid=164, ppid=162, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=f962e4f9abbc7826ddc0ddd4343efcb3, ASSIGN}] 2025-01-08T05:01:29,025 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=162 2025-01-08T05:01:29,027 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=164, ppid=162, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=f962e4f9abbc7826ddc0ddd4343efcb3, ASSIGN 2025-01-08T05:01:29,027 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=163, ppid=162, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=e2b08e9e5645f931a0a5276a05929aba, ASSIGN 2025-01-08T05:01:29,027 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=164, ppid=162, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=f962e4f9abbc7826ddc0ddd4343efcb3, ASSIGN; state=OFFLINE, location=f28094b72178,39801,1736312288779; forceNewPlan=false, retain=false 2025-01-08T05:01:29,027 INFO [PEWorker-1 {}] assignment.TransitRegionStateProcedure(264): Starting pid=163, ppid=162, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=e2b08e9e5645f931a0a5276a05929aba, ASSIGN; state=OFFLINE, location=f28094b72178,37967,1736312288941; forceNewPlan=false, retain=false 2025-01-08T05:01:29,178 INFO [f28094b72178:42195 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T05:01:29,178 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=163 updating hbase:meta row=e2b08e9e5645f931a0a5276a05929aba, regionState=OPENING, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T05:01:29,178 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=164 updating hbase:meta row=f962e4f9abbc7826ddc0ddd4343efcb3, regionState=OPENING, regionLocation=f28094b72178,39801,1736312288779 2025-01-08T05:01:29,180 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=165, ppid=163, state=RUNNABLE; OpenRegionProcedure e2b08e9e5645f931a0a5276a05929aba, server=f28094b72178,37967,1736312288941}] 2025-01-08T05:01:29,180 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=166, ppid=164, state=RUNNABLE; OpenRegionProcedure f962e4f9abbc7826ddc0ddd4343efcb3, server=f28094b72178,39801,1736312288779}] 2025-01-08T05:01:29,226 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=162 2025-01-08T05:01:29,331 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:01:29,332 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,39801,1736312288779 2025-01-08T05:01:29,334 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] handler.AssignRegionHandler(135): Open testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba. 2025-01-08T05:01:29,349 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] handler.AssignRegionHandler(135): Open testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3. 2025-01-08T05:01:29,350 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(7285): Opening region: {ENCODED => f962e4f9abbc7826ddc0ddd4343efcb3, NAME => 'testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T05:01:29,350 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3. service=AccessControlService 2025-01-08T05:01:29,350 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T05:01:29,351 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testEmptyExportFileSystemState f962e4f9abbc7826ddc0ddd4343efcb3 2025-01-08T05:01:29,351 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(894): Instantiated testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:01:29,351 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(7327): checking encryption for f962e4f9abbc7826ddc0ddd4343efcb3 2025-01-08T05:01:29,351 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(7330): checking classloading for f962e4f9abbc7826ddc0ddd4343efcb3 2025-01-08T05:01:29,334 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(7285): Opening region: {ENCODED => e2b08e9e5645f931a0a5276a05929aba, NAME => 'testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T05:01:29,352 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba. service=AccessControlService 2025-01-08T05:01:29,352 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T05:01:29,352 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testEmptyExportFileSystemState e2b08e9e5645f931a0a5276a05929aba 2025-01-08T05:01:29,352 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(894): Instantiated testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:01:29,352 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(7327): checking encryption for e2b08e9e5645f931a0a5276a05929aba 2025-01-08T05:01:29,352 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(7330): checking classloading for e2b08e9e5645f931a0a5276a05929aba 2025-01-08T05:01:29,353 INFO [StoreOpener-f962e4f9abbc7826ddc0ddd4343efcb3-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region f962e4f9abbc7826ddc0ddd4343efcb3 2025-01-08T05:01:29,354 INFO [StoreOpener-e2b08e9e5645f931a0a5276a05929aba-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region e2b08e9e5645f931a0a5276a05929aba 2025-01-08T05:01:29,355 INFO [StoreOpener-f962e4f9abbc7826ddc0ddd4343efcb3-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region f962e4f9abbc7826ddc0ddd4343efcb3 columnFamilyName cf 2025-01-08T05:01:29,355 INFO [StoreOpener-e2b08e9e5645f931a0a5276a05929aba-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region e2b08e9e5645f931a0a5276a05929aba columnFamilyName cf 2025-01-08T05:01:29,356 DEBUG [StoreOpener-f962e4f9abbc7826ddc0ddd4343efcb3-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:01:29,356 DEBUG [StoreOpener-e2b08e9e5645f931a0a5276a05929aba-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:01:29,356 INFO [StoreOpener-f962e4f9abbc7826ddc0ddd4343efcb3-1 {}] regionserver.HStore(327): Store=f962e4f9abbc7826ddc0ddd4343efcb3/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T05:01:29,356 INFO [StoreOpener-e2b08e9e5645f931a0a5276a05929aba-1 {}] regionserver.HStore(327): Store=e2b08e9e5645f931a0a5276a05929aba/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T05:01:29,357 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/f962e4f9abbc7826ddc0ddd4343efcb3 2025-01-08T05:01:29,357 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/e2b08e9e5645f931a0a5276a05929aba 2025-01-08T05:01:29,357 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/f962e4f9abbc7826ddc0ddd4343efcb3 2025-01-08T05:01:29,357 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/e2b08e9e5645f931a0a5276a05929aba 2025-01-08T05:01:29,359 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(1085): writing seq id for e2b08e9e5645f931a0a5276a05929aba 2025-01-08T05:01:29,359 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(1085): writing seq id for f962e4f9abbc7826ddc0ddd4343efcb3 2025-01-08T05:01:29,361 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/e2b08e9e5645f931a0a5276a05929aba/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T05:01:29,361 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/f962e4f9abbc7826ddc0ddd4343efcb3/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T05:01:29,361 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(1102): Opened f962e4f9abbc7826ddc0ddd4343efcb3; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=67663612, jitterRate=0.0082663893699646}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T05:01:29,361 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(1102): Opened e2b08e9e5645f931a0a5276a05929aba; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=67172474, jitterRate=9.478628635406494E-4}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T05:01:29,362 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(1001): Region open journal for e2b08e9e5645f931a0a5276a05929aba: 2025-01-08T05:01:29,362 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(1001): Region open journal for f962e4f9abbc7826ddc0ddd4343efcb3: 2025-01-08T05:01:29,363 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba., pid=165, masterSystemTime=1736312489331 2025-01-08T05:01:29,363 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3., pid=166, masterSystemTime=1736312489332 2025-01-08T05:01:29,364 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3. 2025-01-08T05:01:29,364 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] handler.AssignRegionHandler(164): Opened testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3. 2025-01-08T05:01:29,364 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=164 updating hbase:meta row=f962e4f9abbc7826ddc0ddd4343efcb3, regionState=OPEN, openSeqNum=2, regionLocation=f28094b72178,39801,1736312288779 2025-01-08T05:01:29,364 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba. 2025-01-08T05:01:29,364 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] handler.AssignRegionHandler(164): Opened testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba. 2025-01-08T05:01:29,365 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=163 updating hbase:meta row=e2b08e9e5645f931a0a5276a05929aba, regionState=OPEN, openSeqNum=2, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T05:01:29,367 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=166, resume processing ppid=164 2025-01-08T05:01:29,367 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=165, resume processing ppid=163 2025-01-08T05:01:29,367 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=166, ppid=164, state=SUCCESS; OpenRegionProcedure f962e4f9abbc7826ddc0ddd4343efcb3, server=f28094b72178,39801,1736312288779 in 185 msec 2025-01-08T05:01:29,367 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=165, ppid=163, state=SUCCESS; OpenRegionProcedure e2b08e9e5645f931a0a5276a05929aba, server=f28094b72178,37967,1736312288941 in 187 msec 2025-01-08T05:01:29,368 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=164, ppid=162, state=SUCCESS; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=f962e4f9abbc7826ddc0ddd4343efcb3, ASSIGN in 343 msec 2025-01-08T05:01:29,369 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=163, resume processing ppid=162 2025-01-08T05:01:29,369 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=163, ppid=162, state=SUCCESS; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=e2b08e9e5645f931a0a5276a05929aba, ASSIGN in 343 msec 2025-01-08T05:01:29,369 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=162, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testEmptyExportFileSystemState execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T05:01:29,370 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testEmptyExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312489369"}]},"ts":"1736312489369"} 2025-01-08T05:01:29,371 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testEmptyExportFileSystemState, state=ENABLED in hbase:meta 2025-01-08T05:01:29,373 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=162, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testEmptyExportFileSystemState execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T05:01:29,373 DEBUG [PEWorker-4 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testEmptyExportFileSystemState jenkins: RWXCA 2025-01-08T05:01:29,375 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37583 {}] access.PermissionStorage(611): Read acl: entry[testtb-testEmptyExportFileSystemState], kv [jenkins: RWXCA] 2025-01-08T05:01:29,384 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:29,384 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:29,384 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:29,384 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:29,386 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testEmptyExportFileSystemState with data PBUF\x0AK\x0A\x07jenkins\x12@\x08\x03"<\x0A0\x0A\x07default\x12%testtb-testEmptyExportFileSystemState \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:29,386 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testEmptyExportFileSystemState with data PBUF\x0AK\x0A\x07jenkins\x12@\x08\x03"<\x0A0\x0A\x07default\x12%testtb-testEmptyExportFileSystemState \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:29,386 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testEmptyExportFileSystemState with data PBUF\x0AK\x0A\x07jenkins\x12@\x08\x03"<\x0A0\x0A\x07default\x12%testtb-testEmptyExportFileSystemState \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:29,386 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testEmptyExportFileSystemState with data PBUF\x0AK\x0A\x07jenkins\x12@\x08\x03"<\x0A0\x0A\x07default\x12%testtb-testEmptyExportFileSystemState \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:29,386 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:29,386 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:29,386 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:29,386 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:29,388 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=162, state=SUCCESS; CreateTableProcedure table=testtb-testEmptyExportFileSystemState in 465 msec 2025-01-08T05:01:29,527 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=162 2025-01-08T05:01:29,527 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testEmptyExportFileSystemState, procId: 162 completed 2025-01-08T05:01:29,530 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testEmptyExportFileSystemState 2025-01-08T05:01:29,530 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba. 2025-01-08T05:01:29,530 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T05:01:29,561 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } 2025-01-08T05:01:29,561 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736312489561 (current time:1736312489561). 2025-01-08T05:01:29,561 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T05:01:29,561 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot emptySnaptb0-testEmptyExportFileSystemState VERSION not specified, setting to 2 2025-01-08T05:01:29,561 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T05:01:29,562 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x4a07f8a9 to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@42613a9a 2025-01-08T05:01:29,569 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4c538048, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:01:29,571 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:01:29,572 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:38678, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:01:29,573 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x4a07f8a9 to 127.0.0.1:59048 2025-01-08T05:01:29,573 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:01:29,574 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x7d57e924 to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@36a9b9fe 2025-01-08T05:01:29,584 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@66282ea3, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:01:29,587 DEBUG [hconnection-0x6a950155-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:01:29,588 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:38692, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:01:29,592 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x7d57e924 to 127.0.0.1:59048 2025-01-08T05:01:29,592 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:01:29,592 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] access.PermissionStorage(611): Read acl: entry[testtb-testEmptyExportFileSystemState], kv [jenkins: RWXCA] 2025-01-08T05:01:29,595 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T05:01:29,596 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=167, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } 2025-01-08T05:01:29,596 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 }, snapshot procedure id = 167 2025-01-08T05:01:29,597 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T05:01:29,597 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=167 2025-01-08T05:01:29,600 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T05:01:29,602 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T05:01:29,628 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742225_1401 (size=185) 2025-01-08T05:01:29,630 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742225_1401 (size=185) 2025-01-08T05:01:29,633 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T05:01:29,633 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=168, ppid=167, state=RUNNABLE; SnapshotRegionProcedure e2b08e9e5645f931a0a5276a05929aba}, {pid=169, ppid=167, state=RUNNABLE; SnapshotRegionProcedure f962e4f9abbc7826ddc0ddd4343efcb3}] 2025-01-08T05:01:29,634 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=168, ppid=167, state=RUNNABLE; SnapshotRegionProcedure e2b08e9e5645f931a0a5276a05929aba 2025-01-08T05:01:29,634 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=169, ppid=167, state=RUNNABLE; SnapshotRegionProcedure f962e4f9abbc7826ddc0ddd4343efcb3 2025-01-08T05:01:29,638 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742225_1401 (size=185) 2025-01-08T05:01:29,699 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=167 2025-01-08T05:01:29,785 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:01:29,785 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,39801,1736312288779 2025-01-08T05:01:29,786 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=39801 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=169 2025-01-08T05:01:29,786 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=37967 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=168 2025-01-08T05:01:29,786 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=169}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3. 2025-01-08T05:01:29,786 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=168}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba. 2025-01-08T05:01:29,787 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=168}] regionserver.HRegion(2538): Flush status journal for e2b08e9e5645f931a0a5276a05929aba: 2025-01-08T05:01:29,787 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=169}] regionserver.HRegion(2538): Flush status journal for f962e4f9abbc7826ddc0ddd4343efcb3: 2025-01-08T05:01:29,787 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=168}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba. for emptySnaptb0-testEmptyExportFileSystemState completed. 2025-01-08T05:01:29,787 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=169}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3. for emptySnaptb0-testEmptyExportFileSystemState completed. 2025-01-08T05:01:29,787 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=168}] snapshot.SnapshotManifest(241): Storing 'testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba.' region-info for snapshot=emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T05:01:29,787 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=169}] snapshot.SnapshotManifest(241): Storing 'testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3.' region-info for snapshot=emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T05:01:29,787 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=168}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T05:01:29,787 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=169}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T05:01:29,787 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=168}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T05:01:29,787 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=169}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T05:01:29,816 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742227_1403 (size=76) 2025-01-08T05:01:29,816 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742227_1403 (size=76) 2025-01-08T05:01:29,816 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742227_1403 (size=76) 2025-01-08T05:01:29,817 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=168}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba. 2025-01-08T05:01:29,817 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=168}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=168 2025-01-08T05:01:29,817 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=168 2025-01-08T05:01:29,817 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testEmptyExportFileSystemState on region e2b08e9e5645f931a0a5276a05929aba 2025-01-08T05:01:29,818 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=168, ppid=167, state=RUNNABLE; SnapshotRegionProcedure e2b08e9e5645f931a0a5276a05929aba 2025-01-08T05:01:29,820 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=168, ppid=167, state=SUCCESS; SnapshotRegionProcedure e2b08e9e5645f931a0a5276a05929aba in 186 msec 2025-01-08T05:01:29,825 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742226_1402 (size=76) 2025-01-08T05:01:29,825 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742226_1402 (size=76) 2025-01-08T05:01:29,825 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742226_1402 (size=76) 2025-01-08T05:01:29,900 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=167 2025-01-08T05:01:29,919 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testEmptyExportFileSystemState' 2025-01-08T05:01:30,203 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=167 2025-01-08T05:01:30,229 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=169}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3. 2025-01-08T05:01:30,229 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=169}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=169 2025-01-08T05:01:30,229 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=169 2025-01-08T05:01:30,230 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testEmptyExportFileSystemState on region f962e4f9abbc7826ddc0ddd4343efcb3 2025-01-08T05:01:30,230 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=169, ppid=167, state=RUNNABLE; SnapshotRegionProcedure f962e4f9abbc7826ddc0ddd4343efcb3 2025-01-08T05:01:30,241 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=169, resume processing ppid=167 2025-01-08T05:01:30,241 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T05:01:30,242 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=169, ppid=167, state=SUCCESS; SnapshotRegionProcedure f962e4f9abbc7826ddc0ddd4343efcb3 in 605 msec 2025-01-08T05:01:30,242 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T05:01:30,243 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T05:01:30,243 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T05:01:30,244 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:01:30,244 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(328): No files under family: cf 2025-01-08T05:01:30,270 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742228_1404 (size=68) 2025-01-08T05:01:30,270 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742228_1404 (size=68) 2025-01-08T05:01:30,272 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742228_1404 (size=68) 2025-01-08T05:01:30,273 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T05:01:30,273 DEBUG [PEWorker-5 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T05:01:30,280 DEBUG [PEWorker-5 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T05:01:30,407 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742229_1405 (size=673) 2025-01-08T05:01:30,407 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742229_1405 (size=673) 2025-01-08T05:01:30,407 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742229_1405 (size=673) 2025-01-08T05:01:30,704 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=167 2025-01-08T05:01:30,819 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T05:01:30,846 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T05:01:30,847 DEBUG [PEWorker-5 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/emptySnaptb0-testEmptyExportFileSystemState to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T05:01:30,848 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T05:01:30,849 DEBUG [PEWorker-5 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 }, snapshot procedure id = 167 2025-01-08T05:01:30,850 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=167, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } in 1.2540 sec 2025-01-08T05:01:31,711 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=167 2025-01-08T05:01:31,711 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testEmptyExportFileSystemState, procId: 167 completed 2025-01-08T05:01:31,716 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37967 {}] regionserver.HRegion(8254): writing data to region testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T05:01:31,720 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39801 {}] regionserver.HRegion(8254): writing data to region testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T05:01:31,729 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testEmptyExportFileSystemState 2025-01-08T05:01:31,729 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba. 2025-01-08T05:01:31,729 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T05:01:31,743 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } 2025-01-08T05:01:31,743 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736312491743 (current time:1736312491743). 2025-01-08T05:01:31,743 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T05:01:31,743 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testEmptyExportFileSystemState VERSION not specified, setting to 2 2025-01-08T05:01:31,743 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T05:01:31,744 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x74ae90ab to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@36f3a8c2 2025-01-08T05:01:31,751 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5083a37d, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:01:31,752 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:01:31,753 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:38706, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:01:31,754 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x74ae90ab to 127.0.0.1:59048 2025-01-08T05:01:31,754 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:01:31,755 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x2482518e to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@f1c5864 2025-01-08T05:01:31,762 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@42c98ab1, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:01:31,764 DEBUG [hconnection-0x2bbaeee6-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:01:31,765 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:38712, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:01:31,767 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x2482518e to 127.0.0.1:59048 2025-01-08T05:01:31,767 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:01:31,767 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] access.PermissionStorage(611): Read acl: entry[testtb-testEmptyExportFileSystemState], kv [jenkins: RWXCA] 2025-01-08T05:01:31,769 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T05:01:31,771 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=170, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } 2025-01-08T05:01:31,771 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 }, snapshot procedure id = 170 2025-01-08T05:01:31,772 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T05:01:31,772 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=170 2025-01-08T05:01:31,773 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T05:01:31,775 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T05:01:31,788 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742230_1406 (size=180) 2025-01-08T05:01:31,789 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742230_1406 (size=180) 2025-01-08T05:01:31,789 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742230_1406 (size=180) 2025-01-08T05:01:31,791 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T05:01:31,791 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=171, ppid=170, state=RUNNABLE; SnapshotRegionProcedure e2b08e9e5645f931a0a5276a05929aba}, {pid=172, ppid=170, state=RUNNABLE; SnapshotRegionProcedure f962e4f9abbc7826ddc0ddd4343efcb3}] 2025-01-08T05:01:31,792 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=172, ppid=170, state=RUNNABLE; SnapshotRegionProcedure f962e4f9abbc7826ddc0ddd4343efcb3 2025-01-08T05:01:31,792 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=171, ppid=170, state=RUNNABLE; SnapshotRegionProcedure e2b08e9e5645f931a0a5276a05929aba 2025-01-08T05:01:31,873 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=170 2025-01-08T05:01:31,942 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,39801,1736312288779 2025-01-08T05:01:31,943 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=39801 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=172 2025-01-08T05:01:31,943 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3. 2025-01-08T05:01:31,943 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.HRegion(2837): Flushing f962e4f9abbc7826ddc0ddd4343efcb3 1/1 column families, dataSize=3.06 KB heapSize=6.86 KB 2025-01-08T05:01:31,945 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:01:31,945 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37967 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=171 2025-01-08T05:01:31,945 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba. 2025-01-08T05:01:31,945 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.HRegion(2837): Flushing e2b08e9e5645f931a0a5276a05929aba 1/1 column families, dataSize=199 B heapSize=688 B 2025-01-08T05:01:31,982 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e2025010831f21bb18bb6402cace3a2f1d5ad2ca7_e2b08e9e5645f931a0a5276a05929aba is 71, key is 0232b53cb9f4691d424b8a20dcfa707a/cf:q/1736312491716/Put/seqid=0 2025-01-08T05:01:31,999 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b2025010869a41609937c49b2ba67fcaaef718b52_f962e4f9abbc7826ddc0ddd4343efcb3 is 71, key is 1374135e4481ca300641ac50f04b4a87/cf:q/1736312491720/Put/seqid=0 2025-01-08T05:01:32,028 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742232_1408 (size=8171) 2025-01-08T05:01:32,028 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742232_1408 (size=8171) 2025-01-08T05:01:32,028 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742232_1408 (size=8171) 2025-01-08T05:01:32,029 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:01:32,031 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742231_1407 (size=5101) 2025-01-08T05:01:32,039 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742231_1407 (size=5101) 2025-01-08T05:01:32,044 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742231_1407 (size=5101) 2025-01-08T05:01:32,045 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:01:32,053 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e2025010831f21bb18bb6402cace3a2f1d5ad2ca7_e2b08e9e5645f931a0a5276a05929aba to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/d41d8cd98f00b204e9800998ecf8427e2025010831f21bb18bb6402cace3a2f1d5ad2ca7_e2b08e9e5645f931a0a5276a05929aba 2025-01-08T05:01:32,057 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/e2b08e9e5645f931a0a5276a05929aba/.tmp/cf/7b3c3da82b994c3986a234783891155a, store: [table=testtb-testEmptyExportFileSystemState family=cf region=e2b08e9e5645f931a0a5276a05929aba] 2025-01-08T05:01:32,058 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/e2b08e9e5645f931a0a5276a05929aba/.tmp/cf/7b3c3da82b994c3986a234783891155a is 214, key is 0c4ab2df35b1fedfc2ba9a0c78f8595cc/cf:q/1736312491716/Put/seqid=0 2025-01-08T05:01:32,061 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b2025010869a41609937c49b2ba67fcaaef718b52_f962e4f9abbc7826ddc0ddd4343efcb3 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/c4ca4238a0b923820dcc509a6f75849b2025010869a41609937c49b2ba67fcaaef718b52_f962e4f9abbc7826ddc0ddd4343efcb3 2025-01-08T05:01:32,063 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/f962e4f9abbc7826ddc0ddd4343efcb3/.tmp/cf/7124e7aead5743cfb194ef37599cfbdc, store: [table=testtb-testEmptyExportFileSystemState family=cf region=f962e4f9abbc7826ddc0ddd4343efcb3] 2025-01-08T05:01:32,064 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/f962e4f9abbc7826ddc0ddd4343efcb3/.tmp/cf/7124e7aead5743cfb194ef37599cfbdc is 214, key is 16603984239cee221bd53a9f74dfb7f08/cf:q/1736312491720/Put/seqid=0 2025-01-08T05:01:32,074 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=170 2025-01-08T05:01:32,082 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742233_1409 (size=5936) 2025-01-08T05:01:32,082 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742233_1409 (size=5936) 2025-01-08T05:01:32,082 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742233_1409 (size=5936) 2025-01-08T05:01:32,083 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=199, hasBloomFilter=true, into tmp file hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/e2b08e9e5645f931a0a5276a05929aba/.tmp/cf/7b3c3da82b994c3986a234783891155a 2025-01-08T05:01:32,088 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/e2b08e9e5645f931a0a5276a05929aba/.tmp/cf/7b3c3da82b994c3986a234783891155a as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/e2b08e9e5645f931a0a5276a05929aba/cf/7b3c3da82b994c3986a234783891155a 2025-01-08T05:01:32,093 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/e2b08e9e5645f931a0a5276a05929aba/cf/7b3c3da82b994c3986a234783891155a, entries=3, sequenceid=6, filesize=5.8 K 2025-01-08T05:01:32,094 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.HRegion(3040): Finished flush of dataSize ~199 B/199, heapSize ~672 B/672, currentSize=0 B/0 for e2b08e9e5645f931a0a5276a05929aba in 149ms, sequenceid=6, compaction requested=false 2025-01-08T05:01:32,095 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.HRegion(2538): Flush status journal for e2b08e9e5645f931a0a5276a05929aba: 2025-01-08T05:01:32,095 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba. for snaptb0-testEmptyExportFileSystemState completed. 2025-01-08T05:01:32,095 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] snapshot.SnapshotManifest(241): Storing 'testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba.' region-info for snapshot=snaptb0-testEmptyExportFileSystemState 2025-01-08T05:01:32,095 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T05:01:32,095 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/e2b08e9e5645f931a0a5276a05929aba/cf/7b3c3da82b994c3986a234783891155a] hfiles 2025-01-08T05:01:32,095 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/e2b08e9e5645f931a0a5276a05929aba/cf/7b3c3da82b994c3986a234783891155a for snapshot=snaptb0-testEmptyExportFileSystemState 2025-01-08T05:01:32,116 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742234_1410 (size=15237) 2025-01-08T05:01:32,116 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742234_1410 (size=15237) 2025-01-08T05:01:32,121 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742234_1410 (size=15237) 2025-01-08T05:01:32,126 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=3.1 K, hasBloomFilter=true, into tmp file hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/f962e4f9abbc7826ddc0ddd4343efcb3/.tmp/cf/7124e7aead5743cfb194ef37599cfbdc 2025-01-08T05:01:32,141 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/f962e4f9abbc7826ddc0ddd4343efcb3/.tmp/cf/7124e7aead5743cfb194ef37599cfbdc as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/f962e4f9abbc7826ddc0ddd4343efcb3/cf/7124e7aead5743cfb194ef37599cfbdc 2025-01-08T05:01:32,156 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/f962e4f9abbc7826ddc0ddd4343efcb3/cf/7124e7aead5743cfb194ef37599cfbdc, entries=47, sequenceid=6, filesize=14.9 K 2025-01-08T05:01:32,162 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.HRegion(3040): Finished flush of dataSize ~3.06 KB/3137, heapSize ~6.84 KB/7008, currentSize=0 B/0 for f962e4f9abbc7826ddc0ddd4343efcb3 in 219ms, sequenceid=6, compaction requested=false 2025-01-08T05:01:32,162 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.HRegion(2538): Flush status journal for f962e4f9abbc7826ddc0ddd4343efcb3: 2025-01-08T05:01:32,162 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3. for snaptb0-testEmptyExportFileSystemState completed. 2025-01-08T05:01:32,162 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] snapshot.SnapshotManifest(241): Storing 'testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3.' region-info for snapshot=snaptb0-testEmptyExportFileSystemState 2025-01-08T05:01:32,162 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T05:01:32,162 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/f962e4f9abbc7826ddc0ddd4343efcb3/cf/7124e7aead5743cfb194ef37599cfbdc] hfiles 2025-01-08T05:01:32,162 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/f962e4f9abbc7826ddc0ddd4343efcb3/cf/7124e7aead5743cfb194ef37599cfbdc for snapshot=snaptb0-testEmptyExportFileSystemState 2025-01-08T05:01:32,237 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742235_1411 (size=115) 2025-01-08T05:01:32,241 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742235_1411 (size=115) 2025-01-08T05:01:32,241 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742235_1411 (size=115) 2025-01-08T05:01:32,242 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba. 2025-01-08T05:01:32,242 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=171 2025-01-08T05:01:32,242 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=171 2025-01-08T05:01:32,242 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testEmptyExportFileSystemState on region e2b08e9e5645f931a0a5276a05929aba 2025-01-08T05:01:32,243 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=171, ppid=170, state=RUNNABLE; SnapshotRegionProcedure e2b08e9e5645f931a0a5276a05929aba 2025-01-08T05:01:32,247 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=171, ppid=170, state=SUCCESS; SnapshotRegionProcedure e2b08e9e5645f931a0a5276a05929aba in 454 msec 2025-01-08T05:01:32,288 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742236_1412 (size=115) 2025-01-08T05:01:32,288 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742236_1412 (size=115) 2025-01-08T05:01:32,289 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742236_1412 (size=115) 2025-01-08T05:01:32,289 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3. 2025-01-08T05:01:32,290 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=172 2025-01-08T05:01:32,290 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=172 2025-01-08T05:01:32,290 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testEmptyExportFileSystemState on region f962e4f9abbc7826ddc0ddd4343efcb3 2025-01-08T05:01:32,290 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=172, ppid=170, state=RUNNABLE; SnapshotRegionProcedure f962e4f9abbc7826ddc0ddd4343efcb3 2025-01-08T05:01:32,294 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T05:01:32,295 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=172, resume processing ppid=170 2025-01-08T05:01:32,295 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T05:01:32,296 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=172, ppid=170, state=SUCCESS; SnapshotRegionProcedure f962e4f9abbc7826ddc0ddd4343efcb3 in 501 msec 2025-01-08T05:01:32,297 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T05:01:32,297 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T05:01:32,297 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:01:32,299 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/c4ca4238a0b923820dcc509a6f75849b2025010869a41609937c49b2ba67fcaaef718b52_f962e4f9abbc7826ddc0ddd4343efcb3, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/d41d8cd98f00b204e9800998ecf8427e2025010831f21bb18bb6402cace3a2f1d5ad2ca7_e2b08e9e5645f931a0a5276a05929aba] hfiles 2025-01-08T05:01:32,299 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/c4ca4238a0b923820dcc509a6f75849b2025010869a41609937c49b2ba67fcaaef718b52_f962e4f9abbc7826ddc0ddd4343efcb3 2025-01-08T05:01:32,299 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/d41d8cd98f00b204e9800998ecf8427e2025010831f21bb18bb6402cace3a2f1d5ad2ca7_e2b08e9e5645f931a0a5276a05929aba 2025-01-08T05:01:32,331 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742237_1413 (size=299) 2025-01-08T05:01:32,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742237_1413 (size=299) 2025-01-08T05:01:32,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742237_1413 (size=299) 2025-01-08T05:01:32,375 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=170 2025-01-08T05:01:32,739 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T05:01:32,739 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testEmptyExportFileSystemState 2025-01-08T05:01:32,740 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/snaptb0-testEmptyExportFileSystemState 2025-01-08T05:01:32,843 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742238_1414 (size=983) 2025-01-08T05:01:32,843 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742238_1414 (size=983) 2025-01-08T05:01:32,843 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742238_1414 (size=983) 2025-01-08T05:01:32,876 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=170 2025-01-08T05:01:32,879 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T05:01:32,894 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T05:01:32,894 DEBUG [PEWorker-3 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/snaptb0-testEmptyExportFileSystemState to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testEmptyExportFileSystemState 2025-01-08T05:01:32,903 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T05:01:32,903 DEBUG [PEWorker-3 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 }, snapshot procedure id = 170 2025-01-08T05:01:32,911 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=170, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } in 1.1340 sec 2025-01-08T05:01:33,877 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=170 2025-01-08T05:01:33,878 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testEmptyExportFileSystemState, procId: 170 completed 2025-01-08T05:01:33,878 INFO [Time-limited test {}] snapshot.TestExportSnapshot(468): HDFS export destination path: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312493878 2025-01-08T05:01:33,878 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=hdfs://localhost:45939, tgtDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312493878, rawTgtDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312493878, srcFsUri=hdfs://localhost:45939, srcDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:01:33,926 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:45939, inputRoot=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:01:33,926 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1149249371_22, ugi=jenkins (auth:SIMPLE)]], outputRoot=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312493878, skipTmp=false, initialOutputSnapshotDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312493878/.hbase-snapshot/.tmp/emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T05:01:33,928 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T05:01:33,940 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312493878/.hbase-snapshot/.tmp/emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T05:01:34,028 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742239_1415 (size=673) 2025-01-08T05:01:34,029 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742239_1415 (size=673) 2025-01-08T05:01:34,030 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742239_1415 (size=673) 2025-01-08T05:01:34,033 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742240_1416 (size=185) 2025-01-08T05:01:34,034 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742240_1416 (size=185) 2025-01-08T05:01:34,034 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742240_1416 (size=185) 2025-01-08T05:01:34,037 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.7.0-SNAPSHOT.jar 2025-01-08T05:01:34,037 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T05:01:34,037 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T05:01:34,037 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T05:01:34,174 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T05:01:35,408 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop-1251044135607716919.jar 2025-01-08T05:01:35,409 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T05:01:35,409 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T05:01:35,490 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop-8534132506003709484.jar 2025-01-08T05:01:35,491 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T05:01:35,491 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T05:01:35,491 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T05:01:35,491 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T05:01:35,491 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T05:01:35,492 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T05:01:35,492 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T05:01:35,492 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T05:01:35,492 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T05:01:35,493 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T05:01:35,493 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T05:01:35,493 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T05:01:35,493 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T05:01:35,494 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T05:01:35,494 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T05:01:35,494 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T05:01:35,494 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T05:01:35,494 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T05:01:35,495 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:01:35,495 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:01:35,495 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T05:01:35,495 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:01:35,495 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:01:35,496 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T05:01:35,496 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T05:01:35,563 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742241_1417 (size=127628) 2025-01-08T05:01:35,563 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742241_1417 (size=127628) 2025-01-08T05:01:35,563 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742241_1417 (size=127628) 2025-01-08T05:01:35,586 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742242_1418 (size=2172137) 2025-01-08T05:01:35,586 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742242_1418 (size=2172137) 2025-01-08T05:01:35,586 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742242_1418 (size=2172137) 2025-01-08T05:01:35,596 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742243_1419 (size=213228) 2025-01-08T05:01:35,600 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742243_1419 (size=213228) 2025-01-08T05:01:35,603 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742243_1419 (size=213228) 2025-01-08T05:01:35,625 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742244_1420 (size=1877034) 2025-01-08T05:01:35,625 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742244_1420 (size=1877034) 2025-01-08T05:01:35,625 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742244_1420 (size=1877034) 2025-01-08T05:01:35,684 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742245_1421 (size=6350712) 2025-01-08T05:01:35,684 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742245_1421 (size=6350712) 2025-01-08T05:01:35,685 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742245_1421 (size=6350712) 2025-01-08T05:01:35,791 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T05:01:36,098 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742246_1422 (size=533455) 2025-01-08T05:01:36,098 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742246_1422 (size=533455) 2025-01-08T05:01:36,098 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742246_1422 (size=533455) 2025-01-08T05:01:36,545 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742247_1423 (size=7280644) 2025-01-08T05:01:36,545 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742247_1423 (size=7280644) 2025-01-08T05:01:36,545 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742247_1423 (size=7280644) 2025-01-08T05:01:36,614 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742248_1424 (size=4188619) 2025-01-08T05:01:36,614 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742248_1424 (size=4188619) 2025-01-08T05:01:36,614 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742248_1424 (size=4188619) 2025-01-08T05:01:36,632 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742249_1425 (size=20406) 2025-01-08T05:01:36,632 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742249_1425 (size=20406) 2025-01-08T05:01:36,633 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742249_1425 (size=20406) 2025-01-08T05:01:36,641 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742250_1426 (size=75495) 2025-01-08T05:01:36,641 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742250_1426 (size=75495) 2025-01-08T05:01:36,642 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742250_1426 (size=75495) 2025-01-08T05:01:36,661 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742251_1427 (size=45609) 2025-01-08T05:01:36,662 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742251_1427 (size=45609) 2025-01-08T05:01:36,662 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742251_1427 (size=45609) 2025-01-08T05:01:36,681 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742252_1428 (size=110084) 2025-01-08T05:01:36,681 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742252_1428 (size=110084) 2025-01-08T05:01:36,682 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742252_1428 (size=110084) 2025-01-08T05:01:36,704 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742253_1429 (size=1323991) 2025-01-08T05:01:36,704 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742253_1429 (size=1323991) 2025-01-08T05:01:36,705 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742253_1429 (size=1323991) 2025-01-08T05:01:36,717 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742254_1430 (size=23076) 2025-01-08T05:01:36,717 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742254_1430 (size=23076) 2025-01-08T05:01:36,717 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742254_1430 (size=23076) 2025-01-08T05:01:36,729 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742255_1431 (size=126803) 2025-01-08T05:01:36,729 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742255_1431 (size=126803) 2025-01-08T05:01:36,730 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742255_1431 (size=126803) 2025-01-08T05:01:36,765 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742256_1432 (size=322274) 2025-01-08T05:01:36,766 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742256_1432 (size=322274) 2025-01-08T05:01:36,767 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742256_1432 (size=322274) 2025-01-08T05:01:36,793 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742257_1433 (size=1832290) 2025-01-08T05:01:36,794 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742257_1433 (size=1832290) 2025-01-08T05:01:36,794 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742257_1433 (size=1832290) 2025-01-08T05:01:36,812 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742258_1434 (size=30081) 2025-01-08T05:01:36,812 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742258_1434 (size=30081) 2025-01-08T05:01:36,812 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742258_1434 (size=30081) 2025-01-08T05:01:36,831 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742259_1435 (size=53616) 2025-01-08T05:01:36,832 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742259_1435 (size=53616) 2025-01-08T05:01:36,832 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742259_1435 (size=53616) 2025-01-08T05:01:36,848 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742260_1436 (size=29229) 2025-01-08T05:01:36,848 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742260_1436 (size=29229) 2025-01-08T05:01:36,848 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742260_1436 (size=29229) 2025-01-08T05:01:36,874 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742261_1437 (size=169089) 2025-01-08T05:01:36,874 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742261_1437 (size=169089) 2025-01-08T05:01:36,874 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742261_1437 (size=169089) 2025-01-08T05:01:36,900 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742262_1438 (size=451756) 2025-01-08T05:01:36,900 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742262_1438 (size=451756) 2025-01-08T05:01:36,901 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742262_1438 (size=451756) 2025-01-08T05:01:36,939 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742263_1439 (size=5175431) 2025-01-08T05:01:36,940 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742263_1439 (size=5175431) 2025-01-08T05:01:36,940 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742263_1439 (size=5175431) 2025-01-08T05:01:36,948 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742264_1440 (size=136454) 2025-01-08T05:01:36,948 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742264_1440 (size=136454) 2025-01-08T05:01:36,948 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742264_1440 (size=136454) 2025-01-08T05:01:36,957 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742265_1441 (size=907470) 2025-01-08T05:01:36,958 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742265_1441 (size=907470) 2025-01-08T05:01:36,958 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742265_1441 (size=907470) 2025-01-08T05:01:36,972 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742266_1442 (size=3317408) 2025-01-08T05:01:36,974 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742266_1442 (size=3317408) 2025-01-08T05:01:36,974 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742266_1442 (size=3317408) 2025-01-08T05:01:36,983 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742267_1443 (size=503880) 2025-01-08T05:01:36,983 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742267_1443 (size=503880) 2025-01-08T05:01:36,984 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742267_1443 (size=503880) 2025-01-08T05:01:37,005 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742268_1444 (size=4695811) 2025-01-08T05:01:37,006 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742268_1444 (size=4695811) 2025-01-08T05:01:37,010 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742268_1444 (size=4695811) 2025-01-08T05:01:37,016 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T05:01:37,019 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'emptySnaptb0-testEmptyExportFileSystemState' hfile list 2025-01-08T05:01:37,042 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742269_1445 (size=7) 2025-01-08T05:01:37,043 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742269_1445 (size=7) 2025-01-08T05:01:37,044 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742269_1445 (size=7) 2025-01-08T05:01:37,064 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742270_1446 (size=10) 2025-01-08T05:01:37,064 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742270_1446 (size=10) 2025-01-08T05:01:37,067 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742270_1446 (size=10) 2025-01-08T05:01:37,088 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742271_1447 (size=304782) 2025-01-08T05:01:37,088 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742271_1447 (size=304782) 2025-01-08T05:01:37,089 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742271_1447 (size=304782) 2025-01-08T05:01:37,108 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T05:01:37,108 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T05:01:37,120 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0007_000001 (auth:SIMPLE) from 127.0.0.1:48778 2025-01-08T05:01:38,345 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testEmptyExportFileSystemState 2025-01-08T05:01:38,345 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testEmptyExportFileSystemState Metrics about Tables on a single HBase RegionServer 2025-01-08T05:01:38,345 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportExpiredSnapshot 2025-01-08T05:01:43,848 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T05:01:45,232 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0007_000001 (auth:SIMPLE) from 127.0.0.1:39436 2025-01-08T05:01:45,715 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742272_1448 (size=350432) 2025-01-08T05:01:45,715 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742272_1448 (size=350432) 2025-01-08T05:01:45,715 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742272_1448 (size=350432) 2025-01-08T05:01:46,516 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742273_1449 (size=8568) 2025-01-08T05:01:46,523 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742273_1449 (size=8568) 2025-01-08T05:01:46,523 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742273_1449 (size=8568) 2025-01-08T05:01:46,611 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742274_1450 (size=460) 2025-01-08T05:01:46,616 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742274_1450 (size=460) 2025-01-08T05:01:46,616 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742274_1450 (size=460) 2025-01-08T05:01:46,728 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742275_1451 (size=8568) 2025-01-08T05:01:46,729 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742275_1451 (size=8568) 2025-01-08T05:01:46,731 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742275_1451 (size=8568) 2025-01-08T05:01:46,807 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742276_1452 (size=350432) 2025-01-08T05:01:46,808 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742276_1452 (size=350432) 2025-01-08T05:01:46,808 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742276_1452 (size=350432) 2025-01-08T05:01:48,347 INFO [Time-limited test {}] snapshot.ExportSnapshot(1207): Finalize the Snapshot Export 2025-01-08T05:01:48,349 INFO [Time-limited test {}] snapshot.ExportSnapshot(1218): Verify the exported snapshot's expiration status and integrity. 2025-01-08T05:01:48,386 INFO [Time-limited test {}] snapshot.ExportSnapshot(1224): Export Completed: emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T05:01:48,386 INFO [Time-limited test {}] snapshot.TestExportSnapshot(362): Exported snapshot 2025-01-08T05:01:48,391 INFO [Time-limited test {}] snapshot.TestExportSnapshot(373): Verified filesystem state 2025-01-08T05:01:48,391 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1149249371_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState at hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T05:01:48,395 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState/.snapshotinfo 2025-01-08T05:01:48,395 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState/data.manifest 2025-01-08T05:01:48,395 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1149249371_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312493878/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState at hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312493878/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T05:01:48,399 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312493878/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState/.snapshotinfo 2025-01-08T05:01:48,399 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312493878/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState/data.manifest 2025-01-08T05:01:48,432 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testEmptyExportFileSystemState 2025-01-08T05:01:48,432 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.3 disable testtb-testEmptyExportFileSystemState 2025-01-08T05:01:48,433 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=173, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testEmptyExportFileSystemState 2025-01-08T05:01:48,435 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=173 2025-01-08T05:01:48,435 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testEmptyExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312508435"}]},"ts":"1736312508435"} 2025-01-08T05:01:48,437 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testEmptyExportFileSystemState, state=DISABLING in hbase:meta 2025-01-08T05:01:48,439 INFO [PEWorker-2 {}] procedure.DisableTableProcedure(284): Set testtb-testEmptyExportFileSystemState to state=DISABLING 2025-01-08T05:01:48,440 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=174, ppid=173, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testEmptyExportFileSystemState}] 2025-01-08T05:01:48,442 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=175, ppid=174, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=e2b08e9e5645f931a0a5276a05929aba, UNASSIGN}, {pid=176, ppid=174, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=f962e4f9abbc7826ddc0ddd4343efcb3, UNASSIGN}] 2025-01-08T05:01:48,443 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=176, ppid=174, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=f962e4f9abbc7826ddc0ddd4343efcb3, UNASSIGN 2025-01-08T05:01:48,443 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=175, ppid=174, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=e2b08e9e5645f931a0a5276a05929aba, UNASSIGN 2025-01-08T05:01:48,451 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=176 updating hbase:meta row=f962e4f9abbc7826ddc0ddd4343efcb3, regionState=CLOSING, regionLocation=f28094b72178,39801,1736312288779 2025-01-08T05:01:48,451 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=175 updating hbase:meta row=e2b08e9e5645f931a0a5276a05929aba, regionState=CLOSING, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T05:01:48,453 DEBUG [PEWorker-5 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T05:01:48,454 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=177, ppid=175, state=RUNNABLE; CloseRegionProcedure e2b08e9e5645f931a0a5276a05929aba, server=f28094b72178,37967,1736312288941}] 2025-01-08T05:01:48,454 DEBUG [PEWorker-4 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T05:01:48,457 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=178, ppid=176, state=RUNNABLE; CloseRegionProcedure f962e4f9abbc7826ddc0ddd4343efcb3, server=f28094b72178,39801,1736312288779}] 2025-01-08T05:01:48,536 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=173 2025-01-08T05:01:48,607 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:01:48,608 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] handler.UnassignRegionHandler(124): Close e2b08e9e5645f931a0a5276a05929aba 2025-01-08T05:01:48,608 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T05:01:48,608 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] regionserver.HRegion(1681): Closing e2b08e9e5645f931a0a5276a05929aba, disabling compactions & flushes 2025-01-08T05:01:48,608 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] regionserver.HRegion(1703): Closing region testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba. 2025-01-08T05:01:48,608 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba. 2025-01-08T05:01:48,608 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] regionserver.HRegion(1791): Acquired close lock on testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba. after waiting 0 ms 2025-01-08T05:01:48,608 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] regionserver.HRegion(1801): Updates disabled for region testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba. 2025-01-08T05:01:48,609 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,39801,1736312288779 2025-01-08T05:01:48,609 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] handler.UnassignRegionHandler(124): Close f962e4f9abbc7826ddc0ddd4343efcb3 2025-01-08T05:01:48,610 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T05:01:48,610 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] regionserver.HRegion(1681): Closing f962e4f9abbc7826ddc0ddd4343efcb3, disabling compactions & flushes 2025-01-08T05:01:48,610 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] regionserver.HRegion(1703): Closing region testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3. 2025-01-08T05:01:48,610 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3. 2025-01-08T05:01:48,610 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] regionserver.HRegion(1791): Acquired close lock on testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3. after waiting 0 ms 2025-01-08T05:01:48,610 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] regionserver.HRegion(1801): Updates disabled for region testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3. 2025-01-08T05:01:48,620 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/e2b08e9e5645f931a0a5276a05929aba/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T05:01:48,621 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:01:48,622 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] regionserver.HRegion(1922): Closed testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba. 2025-01-08T05:01:48,622 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] regionserver.HRegion(1635): Region close journal for e2b08e9e5645f931a0a5276a05929aba: 2025-01-08T05:01:48,623 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/f962e4f9abbc7826ddc0ddd4343efcb3/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T05:01:48,624 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:01:48,624 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] regionserver.HRegion(1922): Closed testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3. 2025-01-08T05:01:48,624 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] regionserver.HRegion(1635): Region close journal for f962e4f9abbc7826ddc0ddd4343efcb3: 2025-01-08T05:01:48,625 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] handler.UnassignRegionHandler(170): Closed e2b08e9e5645f931a0a5276a05929aba 2025-01-08T05:01:48,625 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=175 updating hbase:meta row=e2b08e9e5645f931a0a5276a05929aba, regionState=CLOSED 2025-01-08T05:01:48,629 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] handler.UnassignRegionHandler(170): Closed f962e4f9abbc7826ddc0ddd4343efcb3 2025-01-08T05:01:48,630 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=176 updating hbase:meta row=f962e4f9abbc7826ddc0ddd4343efcb3, regionState=CLOSED 2025-01-08T05:01:48,637 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=177, resume processing ppid=175 2025-01-08T05:01:48,637 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=177, ppid=175, state=SUCCESS; CloseRegionProcedure e2b08e9e5645f931a0a5276a05929aba, server=f28094b72178,37967,1736312288941 in 176 msec 2025-01-08T05:01:48,638 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=178, resume processing ppid=176 2025-01-08T05:01:48,638 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=178, ppid=176, state=SUCCESS; CloseRegionProcedure f962e4f9abbc7826ddc0ddd4343efcb3, server=f28094b72178,39801,1736312288779 in 181 msec 2025-01-08T05:01:48,639 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=175, ppid=174, state=SUCCESS; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=e2b08e9e5645f931a0a5276a05929aba, UNASSIGN in 195 msec 2025-01-08T05:01:48,640 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=176, resume processing ppid=174 2025-01-08T05:01:48,640 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=176, ppid=174, state=SUCCESS; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=f962e4f9abbc7826ddc0ddd4343efcb3, UNASSIGN in 196 msec 2025-01-08T05:01:48,644 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=174, resume processing ppid=173 2025-01-08T05:01:48,644 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=174, ppid=173, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testEmptyExportFileSystemState in 201 msec 2025-01-08T05:01:48,646 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testEmptyExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312508646"}]},"ts":"1736312508646"} 2025-01-08T05:01:48,654 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testEmptyExportFileSystemState, state=DISABLED in hbase:meta 2025-01-08T05:01:48,656 INFO [PEWorker-5 {}] procedure.DisableTableProcedure(296): Set testtb-testEmptyExportFileSystemState to state=DISABLED 2025-01-08T05:01:48,658 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=173, state=SUCCESS; DisableTableProcedure table=testtb-testEmptyExportFileSystemState in 224 msec 2025-01-08T05:01:48,737 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=173 2025-01-08T05:01:48,737 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testEmptyExportFileSystemState, procId: 173 completed 2025-01-08T05:01:48,738 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.3 delete testtb-testEmptyExportFileSystemState 2025-01-08T05:01:48,739 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=179, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testEmptyExportFileSystemState 2025-01-08T05:01:48,741 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=179, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testEmptyExportFileSystemState 2025-01-08T05:01:48,741 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testEmptyExportFileSystemState 2025-01-08T05:01:48,742 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=179, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testEmptyExportFileSystemState 2025-01-08T05:01:48,743 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37583 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testEmptyExportFileSystemState 2025-01-08T05:01:48,747 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testEmptyExportFileSystemState 2025-01-08T05:01:48,747 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testEmptyExportFileSystemState 2025-01-08T05:01:48,747 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testEmptyExportFileSystemState 2025-01-08T05:01:48,747 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testEmptyExportFileSystemState 2025-01-08T05:01:48,747 DEBUG [HFileArchiver-19 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/f962e4f9abbc7826ddc0ddd4343efcb3 2025-01-08T05:01:48,748 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testEmptyExportFileSystemState with data PBUF 2025-01-08T05:01:48,748 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testEmptyExportFileSystemState with data PBUF 2025-01-08T05:01:48,747 DEBUG [HFileArchiver-18 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/e2b08e9e5645f931a0a5276a05929aba 2025-01-08T05:01:48,748 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testEmptyExportFileSystemState with data PBUF 2025-01-08T05:01:48,749 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testEmptyExportFileSystemState with data PBUF 2025-01-08T05:01:48,750 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testEmptyExportFileSystemState 2025-01-08T05:01:48,750 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:48,750 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testEmptyExportFileSystemState 2025-01-08T05:01:48,750 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:48,750 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testEmptyExportFileSystemState 2025-01-08T05:01:48,750 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:48,750 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testEmptyExportFileSystemState 2025-01-08T05:01:48,750 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:48,751 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:48,753 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:48,753 DEBUG [HFileArchiver-19 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/f962e4f9abbc7826ddc0ddd4343efcb3/cf, FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/f962e4f9abbc7826ddc0ddd4343efcb3/recovered.edits] 2025-01-08T05:01:48,753 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=179 2025-01-08T05:01:48,753 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:48,754 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:48,754 DEBUG [HFileArchiver-18 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/e2b08e9e5645f931a0a5276a05929aba/cf, FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/e2b08e9e5645f931a0a5276a05929aba/recovered.edits] 2025-01-08T05:01:48,764 DEBUG [HFileArchiver-19 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/f962e4f9abbc7826ddc0ddd4343efcb3/cf/7124e7aead5743cfb194ef37599cfbdc to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testEmptyExportFileSystemState/f962e4f9abbc7826ddc0ddd4343efcb3/cf/7124e7aead5743cfb194ef37599cfbdc 2025-01-08T05:01:48,764 DEBUG [HFileArchiver-18 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/e2b08e9e5645f931a0a5276a05929aba/cf/7b3c3da82b994c3986a234783891155a to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testEmptyExportFileSystemState/e2b08e9e5645f931a0a5276a05929aba/cf/7b3c3da82b994c3986a234783891155a 2025-01-08T05:01:48,770 DEBUG [HFileArchiver-18 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/e2b08e9e5645f931a0a5276a05929aba/recovered.edits/9.seqid to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testEmptyExportFileSystemState/e2b08e9e5645f931a0a5276a05929aba/recovered.edits/9.seqid 2025-01-08T05:01:48,771 DEBUG [HFileArchiver-19 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/f962e4f9abbc7826ddc0ddd4343efcb3/recovered.edits/9.seqid to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testEmptyExportFileSystemState/f962e4f9abbc7826ddc0ddd4343efcb3/recovered.edits/9.seqid 2025-01-08T05:01:48,771 DEBUG [HFileArchiver-18 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/e2b08e9e5645f931a0a5276a05929aba 2025-01-08T05:01:48,772 DEBUG [HFileArchiver-19 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testEmptyExportFileSystemState/f962e4f9abbc7826ddc0ddd4343efcb3 2025-01-08T05:01:48,772 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(313): Archived testtb-testEmptyExportFileSystemState regions 2025-01-08T05:01:48,773 DEBUG [PEWorker-3 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0 2025-01-08T05:01:48,774 DEBUG [PEWorker-3 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf] 2025-01-08T05:01:48,781 DEBUG [PEWorker-3 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/c4ca4238a0b923820dcc509a6f75849b2025010869a41609937c49b2ba67fcaaef718b52_f962e4f9abbc7826ddc0ddd4343efcb3 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/c4ca4238a0b923820dcc509a6f75849b2025010869a41609937c49b2ba67fcaaef718b52_f962e4f9abbc7826ddc0ddd4343efcb3 2025-01-08T05:01:48,787 DEBUG [PEWorker-3 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/d41d8cd98f00b204e9800998ecf8427e2025010831f21bb18bb6402cace3a2f1d5ad2ca7_e2b08e9e5645f931a0a5276a05929aba to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/d41d8cd98f00b204e9800998ecf8427e2025010831f21bb18bb6402cace3a2f1d5ad2ca7_e2b08e9e5645f931a0a5276a05929aba 2025-01-08T05:01:48,788 DEBUG [PEWorker-3 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0 2025-01-08T05:01:48,794 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=179, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testEmptyExportFileSystemState 2025-01-08T05:01:48,797 WARN [PEWorker-3 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testtb-testEmptyExportFileSystemState from hbase:meta 2025-01-08T05:01:48,825 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testEmptyExportFileSystemState' descriptor. 2025-01-08T05:01:48,828 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=179, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testEmptyExportFileSystemState 2025-01-08T05:01:48,828 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testEmptyExportFileSystemState' from region states. 2025-01-08T05:01:48,828 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736312508828"}]},"ts":"9223372036854775807"} 2025-01-08T05:01:48,828 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736312508828"}]},"ts":"9223372036854775807"} 2025-01-08T05:01:48,853 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T05:01:48,853 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => e2b08e9e5645f931a0a5276a05929aba, NAME => 'testtb-testEmptyExportFileSystemState,,1736312488920.e2b08e9e5645f931a0a5276a05929aba.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => f962e4f9abbc7826ddc0ddd4343efcb3, NAME => 'testtb-testEmptyExportFileSystemState,1,1736312488920.f962e4f9abbc7826ddc0ddd4343efcb3.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T05:01:48,853 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testEmptyExportFileSystemState' as deleted. 2025-01-08T05:01:48,853 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testEmptyExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736312508853"}]},"ts":"9223372036854775807"} 2025-01-08T05:01:48,855 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=179 2025-01-08T05:01:48,862 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testEmptyExportFileSystemState state from META 2025-01-08T05:01:48,865 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(133): Finished pid=179, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testEmptyExportFileSystemState 2025-01-08T05:01:48,867 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=179, state=SUCCESS; DeleteTableProcedure table=testtb-testEmptyExportFileSystemState in 127 msec 2025-01-08T05:01:49,056 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=179 2025-01-08T05:01:49,056 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testEmptyExportFileSystemState, procId: 179 completed 2025-01-08T05:01:49,065 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "emptySnaptb0-testEmptyExportFileSystemState" 2025-01-08T05:01:49,068 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(380): Deleting snapshot: emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T05:01:49,070 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "snaptb0-testEmptyExportFileSystemState" 2025-01-08T05:01:49,072 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testEmptyExportFileSystemState 2025-01-08T05:01:49,103 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: snapshot.TestMobSecureExportSnapshot#testEmptyExportFileSystemState Thread=803 (was 796) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:43089 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x168697cd-shared-pool-40 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-19 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x168697cd-shared-pool-42 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-18 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x168697cd-shared-pool-39 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (366384994) connection to localhost/127.0.0.1:43089 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: ApplicationMasterLauncher #12 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1149249371_22 at /127.0.0.1:55998 [Waiting for operation #3] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Thread-6013 java.base@17.0.11/java.io.FileInputStream.readBytes(Native Method) java.base@17.0.11/java.io.FileInputStream.read(FileInputStream.java:276) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:282) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) java.base@17.0.11/sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:281) java.base@17.0.11/sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:324) java.base@17.0.11/sun.nio.cs.StreamDecoder.read(StreamDecoder.java:189) java.base@17.0.11/java.io.InputStreamReader.read(InputStreamReader.java:177) java.base@17.0.11/java.io.BufferedReader.fill(BufferedReader.java:162) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:329) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:396) app//org.apache.hadoop.util.Shell$1.run(Shell.java:1025) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1149249371_22 at /127.0.0.1:54704 [Waiting for operation #4] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: process reaper (pid 25328) java.base@17.0.11/java.lang.ProcessHandleImpl.waitForProcessExit0(Native Method) java.base@17.0.11/java.lang.ProcessHandleImpl$1.run(ProcessHandleImpl.java:150) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1149249371_22 at /127.0.0.1:60670 [Waiting for operation #5] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1465691236_1 at /127.0.0.1:55962 [Waiting for operation #3] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x168697cd-shared-pool-41 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) - Thread LEAK? -, OpenFileDescriptor=808 (was 797) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=931 (was 882) - SystemLoadAverage LEAK? -, ProcessCount=16 (was 16), AvailableMemoryMB=1374 (was 2100) 2025-01-08T05:01:49,103 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=803 is superior to 500 2025-01-08T05:01:49,131 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: snapshot.TestMobSecureExportSnapshot#testExportWithChecksum Thread=803, OpenFileDescriptor=806, MaxFileDescriptor=1048576, SystemLoadAverage=931, ProcessCount=16, AvailableMemoryMB=1373 2025-01-08T05:01:49,131 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=803 is superior to 500 2025-01-08T05:01:49,133 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.3 create 'testtb-testExportWithChecksum', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T05:01:49,135 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=180, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testExportWithChecksum 2025-01-08T05:01:49,136 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=180, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportWithChecksum execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T05:01:49,136 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "testtb-testExportWithChecksum" procId is: 180 2025-01-08T05:01:49,137 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=180, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testExportWithChecksum execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T05:01:49,137 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=180 2025-01-08T05:01:49,227 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742277_1453 (size=440) 2025-01-08T05:01:49,227 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742277_1453 (size=440) 2025-01-08T05:01:49,228 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742277_1453 (size=440) 2025-01-08T05:01:49,232 INFO [RegionOpenAndInit-testtb-testExportWithChecksum-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => 150c2aaa34b578895b86a687c341c2e5, NAME => 'testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testtb-testExportWithChecksum', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:01:49,237 INFO [RegionOpenAndInit-testtb-testExportWithChecksum-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 0a15cae7e53ceb637248f515514255ec, NAME => 'testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testtb-testExportWithChecksum', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:01:49,239 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=180 2025-01-08T05:01:49,268 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742278_1454 (size=65) 2025-01-08T05:01:49,269 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742278_1454 (size=65) 2025-01-08T05:01:49,271 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742278_1454 (size=65) 2025-01-08T05:01:49,272 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742279_1455 (size=65) 2025-01-08T05:01:49,273 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:01:49,273 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-1 {}] regionserver.HRegion(1681): Closing 150c2aaa34b578895b86a687c341c2e5, disabling compactions & flushes 2025-01-08T05:01:49,273 INFO [RegionOpenAndInit-testtb-testExportWithChecksum-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5. 2025-01-08T05:01:49,273 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5. 2025-01-08T05:01:49,273 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5. after waiting 0 ms 2025-01-08T05:01:49,273 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5. 2025-01-08T05:01:49,273 INFO [RegionOpenAndInit-testtb-testExportWithChecksum-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5. 2025-01-08T05:01:49,273 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-1 {}] regionserver.HRegion(1635): Region close journal for 150c2aaa34b578895b86a687c341c2e5: 2025-01-08T05:01:49,273 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:01:49,273 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-0 {}] regionserver.HRegion(1681): Closing 0a15cae7e53ceb637248f515514255ec, disabling compactions & flushes 2025-01-08T05:01:49,274 INFO [RegionOpenAndInit-testtb-testExportWithChecksum-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec. 2025-01-08T05:01:49,274 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec. 2025-01-08T05:01:49,274 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec. after waiting 0 ms 2025-01-08T05:01:49,274 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec. 2025-01-08T05:01:49,274 INFO [RegionOpenAndInit-testtb-testExportWithChecksum-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec. 2025-01-08T05:01:49,274 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-0 {}] regionserver.HRegion(1635): Region close journal for 0a15cae7e53ceb637248f515514255ec: 2025-01-08T05:01:49,274 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742279_1455 (size=65) 2025-01-08T05:01:49,274 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742279_1455 (size=65) 2025-01-08T05:01:49,275 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=180, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testExportWithChecksum execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T05:01:49,275 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5.","families":{"info":[{"qualifier":"regioninfo","vlen":64,"tag":[],"timestamp":"1736312509275"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736312509275"}]},"ts":"1736312509275"} 2025-01-08T05:01:49,275 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec.","families":{"info":[{"qualifier":"regioninfo","vlen":64,"tag":[],"timestamp":"1736312509275"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736312509275"}]},"ts":"1736312509275"} 2025-01-08T05:01:49,277 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T05:01:49,277 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=180, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testExportWithChecksum execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T05:01:49,277 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithChecksum","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312509277"}]},"ts":"1736312509277"} 2025-01-08T05:01:49,278 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithChecksum, state=ENABLING in hbase:meta 2025-01-08T05:01:49,283 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(202): Hosts are {f28094b72178=0} racks are {/default-rack=0} 2025-01-08T05:01:49,284 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T05:01:49,284 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T05:01:49,284 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T05:01:49,284 INFO [PEWorker-1 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T05:01:49,284 INFO [PEWorker-1 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T05:01:49,284 INFO [PEWorker-1 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T05:01:49,284 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T05:01:49,284 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=181, ppid=180, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=0a15cae7e53ceb637248f515514255ec, ASSIGN}, {pid=182, ppid=180, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=150c2aaa34b578895b86a687c341c2e5, ASSIGN}] 2025-01-08T05:01:49,287 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=182, ppid=180, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=150c2aaa34b578895b86a687c341c2e5, ASSIGN 2025-01-08T05:01:49,287 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=181, ppid=180, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=0a15cae7e53ceb637248f515514255ec, ASSIGN 2025-01-08T05:01:49,287 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=182, ppid=180, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=150c2aaa34b578895b86a687c341c2e5, ASSIGN; state=OFFLINE, location=f28094b72178,37967,1736312288941; forceNewPlan=false, retain=false 2025-01-08T05:01:49,287 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=181, ppid=180, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=0a15cae7e53ceb637248f515514255ec, ASSIGN; state=OFFLINE, location=f28094b72178,37583,1736312288873; forceNewPlan=false, retain=false 2025-01-08T05:01:49,438 INFO [f28094b72178:42195 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T05:01:49,438 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=182 updating hbase:meta row=150c2aaa34b578895b86a687c341c2e5, regionState=OPENING, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T05:01:49,438 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=181 updating hbase:meta row=0a15cae7e53ceb637248f515514255ec, regionState=OPENING, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T05:01:49,439 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=183, ppid=181, state=RUNNABLE; OpenRegionProcedure 0a15cae7e53ceb637248f515514255ec, server=f28094b72178,37583,1736312288873}] 2025-01-08T05:01:49,440 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=180 2025-01-08T05:01:49,440 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=184, ppid=182, state=RUNNABLE; OpenRegionProcedure 150c2aaa34b578895b86a687c341c2e5, server=f28094b72178,37967,1736312288941}] 2025-01-08T05:01:49,591 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T05:01:49,591 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:01:49,598 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] handler.AssignRegionHandler(135): Open testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec. 2025-01-08T05:01:49,598 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(7285): Opening region: {ENCODED => 0a15cae7e53ceb637248f515514255ec, NAME => 'testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T05:01:49,599 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec. service=AccessControlService 2025-01-08T05:01:49,599 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T05:01:49,599 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportWithChecksum 0a15cae7e53ceb637248f515514255ec 2025-01-08T05:01:49,599 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(894): Instantiated testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:01:49,599 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] handler.AssignRegionHandler(135): Open testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5. 2025-01-08T05:01:49,599 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(7327): checking encryption for 0a15cae7e53ceb637248f515514255ec 2025-01-08T05:01:49,600 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(7330): checking classloading for 0a15cae7e53ceb637248f515514255ec 2025-01-08T05:01:49,600 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(7285): Opening region: {ENCODED => 150c2aaa34b578895b86a687c341c2e5, NAME => 'testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T05:01:49,600 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5. service=AccessControlService 2025-01-08T05:01:49,600 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T05:01:49,600 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportWithChecksum 150c2aaa34b578895b86a687c341c2e5 2025-01-08T05:01:49,600 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(894): Instantiated testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:01:49,600 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(7327): checking encryption for 150c2aaa34b578895b86a687c341c2e5 2025-01-08T05:01:49,600 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(7330): checking classloading for 150c2aaa34b578895b86a687c341c2e5 2025-01-08T05:01:49,601 INFO [StoreOpener-0a15cae7e53ceb637248f515514255ec-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 0a15cae7e53ceb637248f515514255ec 2025-01-08T05:01:49,602 INFO [StoreOpener-150c2aaa34b578895b86a687c341c2e5-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 150c2aaa34b578895b86a687c341c2e5 2025-01-08T05:01:49,603 INFO [StoreOpener-0a15cae7e53ceb637248f515514255ec-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 0a15cae7e53ceb637248f515514255ec columnFamilyName cf 2025-01-08T05:01:49,604 INFO [StoreOpener-150c2aaa34b578895b86a687c341c2e5-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 150c2aaa34b578895b86a687c341c2e5 columnFamilyName cf 2025-01-08T05:01:49,604 DEBUG [StoreOpener-0a15cae7e53ceb637248f515514255ec-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:01:49,604 DEBUG [StoreOpener-150c2aaa34b578895b86a687c341c2e5-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:01:49,604 INFO [StoreOpener-0a15cae7e53ceb637248f515514255ec-1 {}] regionserver.HStore(327): Store=0a15cae7e53ceb637248f515514255ec/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T05:01:49,605 INFO [StoreOpener-150c2aaa34b578895b86a687c341c2e5-1 {}] regionserver.HStore(327): Store=150c2aaa34b578895b86a687c341c2e5/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T05:01:49,606 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/0a15cae7e53ceb637248f515514255ec 2025-01-08T05:01:49,606 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/150c2aaa34b578895b86a687c341c2e5 2025-01-08T05:01:49,607 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/150c2aaa34b578895b86a687c341c2e5 2025-01-08T05:01:49,607 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/0a15cae7e53ceb637248f515514255ec 2025-01-08T05:01:49,609 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(1085): writing seq id for 0a15cae7e53ceb637248f515514255ec 2025-01-08T05:01:49,610 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(1085): writing seq id for 150c2aaa34b578895b86a687c341c2e5 2025-01-08T05:01:49,616 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/150c2aaa34b578895b86a687c341c2e5/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T05:01:49,616 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/0a15cae7e53ceb637248f515514255ec/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T05:01:49,616 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(1102): Opened 150c2aaa34b578895b86a687c341c2e5; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=68892547, jitterRate=0.02657894790172577}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T05:01:49,616 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(1102): Opened 0a15cae7e53ceb637248f515514255ec; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=59130346, jitterRate=-0.11888918280601501}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T05:01:49,617 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(1001): Region open journal for 150c2aaa34b578895b86a687c341c2e5: 2025-01-08T05:01:49,617 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(1001): Region open journal for 0a15cae7e53ceb637248f515514255ec: 2025-01-08T05:01:49,618 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec., pid=183, masterSystemTime=1736312509591 2025-01-08T05:01:49,618 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5., pid=184, masterSystemTime=1736312509591 2025-01-08T05:01:49,620 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec. 2025-01-08T05:01:49,620 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] handler.AssignRegionHandler(164): Opened testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec. 2025-01-08T05:01:49,620 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=181 updating hbase:meta row=0a15cae7e53ceb637248f515514255ec, regionState=OPEN, openSeqNum=2, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T05:01:49,620 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5. 2025-01-08T05:01:49,620 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] handler.AssignRegionHandler(164): Opened testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5. 2025-01-08T05:01:49,621 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=182 updating hbase:meta row=150c2aaa34b578895b86a687c341c2e5, regionState=OPEN, openSeqNum=2, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T05:01:49,626 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=183, resume processing ppid=181 2025-01-08T05:01:49,626 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=183, ppid=181, state=SUCCESS; OpenRegionProcedure 0a15cae7e53ceb637248f515514255ec, server=f28094b72178,37583,1736312288873 in 182 msec 2025-01-08T05:01:49,627 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=184, resume processing ppid=182 2025-01-08T05:01:49,627 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=184, ppid=182, state=SUCCESS; OpenRegionProcedure 150c2aaa34b578895b86a687c341c2e5, server=f28094b72178,37967,1736312288941 in 183 msec 2025-01-08T05:01:49,628 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=181, ppid=180, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=0a15cae7e53ceb637248f515514255ec, ASSIGN in 342 msec 2025-01-08T05:01:49,629 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=182, resume processing ppid=180 2025-01-08T05:01:49,629 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=182, ppid=180, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=150c2aaa34b578895b86a687c341c2e5, ASSIGN in 343 msec 2025-01-08T05:01:49,630 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=180, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testExportWithChecksum execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T05:01:49,630 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithChecksum","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312509630"}]},"ts":"1736312509630"} 2025-01-08T05:01:49,632 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithChecksum, state=ENABLED in hbase:meta 2025-01-08T05:01:49,634 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=180, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportWithChecksum execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T05:01:49,634 DEBUG [PEWorker-4 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testExportWithChecksum jenkins: RWXCA 2025-01-08T05:01:49,636 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37583 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportWithChecksum], kv [jenkins: RWXCA] 2025-01-08T05:01:49,638 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:49,638 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:49,638 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:49,638 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:01:49,640 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:49,641 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithChecksum with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithChecksum \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:49,641 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:49,641 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:49,641 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithChecksum with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithChecksum \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:49,641 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithChecksum with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithChecksum \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:49,641 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:49,641 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithChecksum with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithChecksum \x00 \x01 \x02 \x03 \x04 2025-01-08T05:01:49,643 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=180, state=SUCCESS; CreateTableProcedure table=testtb-testExportWithChecksum in 507 msec 2025-01-08T05:01:49,740 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=180 2025-01-08T05:01:49,741 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testExportWithChecksum, procId: 180 completed 2025-01-08T05:01:49,743 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportWithChecksum 2025-01-08T05:01:49,743 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec. 2025-01-08T05:01:49,743 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T05:01:49,755 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } 2025-01-08T05:01:49,755 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736312509755 (current time:1736312509755). 2025-01-08T05:01:49,755 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T05:01:49,755 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot emptySnaptb0-testExportWithChecksum VERSION not specified, setting to 2 2025-01-08T05:01:49,755 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T05:01:49,756 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x1cdc4ed9 to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@fada458 2025-01-08T05:01:49,759 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@3da288cf, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:01:49,760 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:01:49,762 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:56630, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:01:49,763 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x1cdc4ed9 to 127.0.0.1:59048 2025-01-08T05:01:49,763 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:01:49,764 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x54d5267b to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@36b8d7dd 2025-01-08T05:01:49,775 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1f75c535, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:01:49,776 DEBUG [hconnection-0x3dad9300-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:01:49,777 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:56644, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:01:49,779 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x54d5267b to 127.0.0.1:59048 2025-01-08T05:01:49,780 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:01:49,780 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportWithChecksum], kv [jenkins: RWXCA] 2025-01-08T05:01:49,780 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T05:01:49,781 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=185, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } 2025-01-08T05:01:49,781 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 }, snapshot procedure id = 185 2025-01-08T05:01:49,782 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=185 2025-01-08T05:01:49,782 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T05:01:49,783 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T05:01:49,784 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T05:01:49,789 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742280_1456 (size=161) 2025-01-08T05:01:49,789 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742280_1456 (size=161) 2025-01-08T05:01:49,790 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742280_1456 (size=161) 2025-01-08T05:01:49,791 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T05:01:49,791 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=186, ppid=185, state=RUNNABLE; SnapshotRegionProcedure 0a15cae7e53ceb637248f515514255ec}, {pid=187, ppid=185, state=RUNNABLE; SnapshotRegionProcedure 150c2aaa34b578895b86a687c341c2e5}] 2025-01-08T05:01:49,791 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=186, ppid=185, state=RUNNABLE; SnapshotRegionProcedure 0a15cae7e53ceb637248f515514255ec 2025-01-08T05:01:49,791 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=187, ppid=185, state=RUNNABLE; SnapshotRegionProcedure 150c2aaa34b578895b86a687c341c2e5 2025-01-08T05:01:49,883 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=185 2025-01-08T05:01:49,942 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T05:01:49,942 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:01:49,943 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37583 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=186 2025-01-08T05:01:49,943 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=37967 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=187 2025-01-08T05:01:49,943 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=186}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec. 2025-01-08T05:01:49,943 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=187}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5. 2025-01-08T05:01:49,943 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=186}] regionserver.HRegion(2538): Flush status journal for 0a15cae7e53ceb637248f515514255ec: 2025-01-08T05:01:49,943 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=186}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec. for emptySnaptb0-testExportWithChecksum completed. 2025-01-08T05:01:49,943 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=187}] regionserver.HRegion(2538): Flush status journal for 150c2aaa34b578895b86a687c341c2e5: 2025-01-08T05:01:49,943 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=187}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5. for emptySnaptb0-testExportWithChecksum completed. 2025-01-08T05:01:49,943 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=186}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec.' region-info for snapshot=emptySnaptb0-testExportWithChecksum 2025-01-08T05:01:49,943 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=186}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T05:01:49,943 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=186}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T05:01:49,943 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=187}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5.' region-info for snapshot=emptySnaptb0-testExportWithChecksum 2025-01-08T05:01:49,943 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=187}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T05:01:49,943 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=187}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T05:01:49,954 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742282_1458 (size=68) 2025-01-08T05:01:49,955 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742282_1458 (size=68) 2025-01-08T05:01:49,956 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742281_1457 (size=68) 2025-01-08T05:01:49,956 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742281_1457 (size=68) 2025-01-08T05:01:49,956 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=186}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec. 2025-01-08T05:01:49,956 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=186}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=186 2025-01-08T05:01:49,956 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742281_1457 (size=68) 2025-01-08T05:01:49,956 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=187}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5. 2025-01-08T05:01:49,956 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742282_1458 (size=68) 2025-01-08T05:01:49,956 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=187}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=187 2025-01-08T05:01:49,956 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=186 2025-01-08T05:01:49,957 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportWithChecksum on region 0a15cae7e53ceb637248f515514255ec 2025-01-08T05:01:49,957 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=187 2025-01-08T05:01:49,957 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportWithChecksum on region 150c2aaa34b578895b86a687c341c2e5 2025-01-08T05:01:49,957 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=186, ppid=185, state=RUNNABLE; SnapshotRegionProcedure 0a15cae7e53ceb637248f515514255ec 2025-01-08T05:01:49,957 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=187, ppid=185, state=RUNNABLE; SnapshotRegionProcedure 150c2aaa34b578895b86a687c341c2e5 2025-01-08T05:01:49,959 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=186, ppid=185, state=SUCCESS; SnapshotRegionProcedure 0a15cae7e53ceb637248f515514255ec in 167 msec 2025-01-08T05:01:49,959 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=187, resume processing ppid=185 2025-01-08T05:01:49,959 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=187, ppid=185, state=SUCCESS; SnapshotRegionProcedure 150c2aaa34b578895b86a687c341c2e5 in 167 msec 2025-01-08T05:01:49,959 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T05:01:49,960 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T05:01:49,961 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T05:01:49,961 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T05:01:49,961 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:01:49,962 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(328): No files under family: cf 2025-01-08T05:01:49,967 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742283_1459 (size=60) 2025-01-08T05:01:49,967 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742283_1459 (size=60) 2025-01-08T05:01:49,968 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742283_1459 (size=60) 2025-01-08T05:01:49,969 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T05:01:49,969 DEBUG [PEWorker-2 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for emptySnaptb0-testExportWithChecksum 2025-01-08T05:01:49,970 DEBUG [PEWorker-2 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/emptySnaptb0-testExportWithChecksum 2025-01-08T05:01:49,979 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742284_1460 (size=641) 2025-01-08T05:01:49,979 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742284_1460 (size=641) 2025-01-08T05:01:49,980 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742284_1460 (size=641) 2025-01-08T05:01:49,982 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T05:01:49,986 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T05:01:49,986 DEBUG [PEWorker-2 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/emptySnaptb0-testExportWithChecksum to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/emptySnaptb0-testExportWithChecksum 2025-01-08T05:01:49,987 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T05:01:49,987 DEBUG [PEWorker-2 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 }, snapshot procedure id = 185 2025-01-08T05:01:49,988 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=185, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } in 207 msec 2025-01-08T05:01:50,084 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=185 2025-01-08T05:01:50,084 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportWithChecksum, procId: 185 completed 2025-01-08T05:01:50,087 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37583 {}] regionserver.HRegion(8254): writing data to region testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T05:01:50,088 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37967 {}] regionserver.HRegion(8254): writing data to region testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T05:01:50,091 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportWithChecksum 2025-01-08T05:01:50,091 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec. 2025-01-08T05:01:50,092 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T05:01:50,103 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } 2025-01-08T05:01:50,103 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736312510103 (current time:1736312510103). 2025-01-08T05:01:50,103 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T05:01:50,103 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testExportWithChecksum VERSION not specified, setting to 2 2025-01-08T05:01:50,103 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T05:01:50,104 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x28b772c4 to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@583ed61b 2025-01-08T05:01:50,119 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@27f16fd7, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:01:50,121 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:01:50,122 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:56656, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:01:50,123 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x28b772c4 to 127.0.0.1:59048 2025-01-08T05:01:50,123 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:01:50,124 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x71e2661e to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@5fc2f837 2025-01-08T05:01:50,131 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7361abcf, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:01:50,133 DEBUG [hconnection-0x119da2e8-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:01:50,134 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:56670, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:01:50,136 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x71e2661e to 127.0.0.1:59048 2025-01-08T05:01:50,136 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:01:50,136 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportWithChecksum], kv [jenkins: RWXCA] 2025-01-08T05:01:50,137 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T05:01:50,138 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=188, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } 2025-01-08T05:01:50,138 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 }, snapshot procedure id = 188 2025-01-08T05:01:50,138 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T05:01:50,139 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=188 2025-01-08T05:01:50,139 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T05:01:50,142 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T05:01:50,155 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742285_1461 (size=156) 2025-01-08T05:01:50,155 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742285_1461 (size=156) 2025-01-08T05:01:50,155 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742285_1461 (size=156) 2025-01-08T05:01:50,158 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T05:01:50,158 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=189, ppid=188, state=RUNNABLE; SnapshotRegionProcedure 0a15cae7e53ceb637248f515514255ec}, {pid=190, ppid=188, state=RUNNABLE; SnapshotRegionProcedure 150c2aaa34b578895b86a687c341c2e5}] 2025-01-08T05:01:50,159 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=189, ppid=188, state=RUNNABLE; SnapshotRegionProcedure 0a15cae7e53ceb637248f515514255ec 2025-01-08T05:01:50,159 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=190, ppid=188, state=RUNNABLE; SnapshotRegionProcedure 150c2aaa34b578895b86a687c341c2e5 2025-01-08T05:01:50,239 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=188 2025-01-08T05:01:50,310 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T05:01:50,310 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:01:50,310 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37967 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=190 2025-01-08T05:01:50,310 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37583 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=189 2025-01-08T05:01:50,311 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec. 2025-01-08T05:01:50,311 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5. 2025-01-08T05:01:50,311 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.HRegion(2837): Flushing 0a15cae7e53ceb637248f515514255ec 1/1 column families, dataSize=266 B heapSize=832 B 2025-01-08T05:01:50,311 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.HRegion(2837): Flushing 150c2aaa34b578895b86a687c341c2e5 1/1 column families, dataSize=3.00 KB heapSize=6.72 KB 2025-01-08T05:01:50,329 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108d2bb5b0ab69241d7822a30d327032e54_150c2aaa34b578895b86a687c341c2e5 is 71, key is 16cb6a3e6491516c1dbf0e24c97dc60b/cf:q/1736312510088/Put/seqid=0 2025-01-08T05:01:50,334 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108cf3c5da2171f408b852005f2c74f7f44_0a15cae7e53ceb637248f515514255ec is 71, key is 045bfb555c2cad068147d92a55337d72/cf:q/1736312510087/Put/seqid=0 2025-01-08T05:01:50,348 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742286_1462 (size=8101) 2025-01-08T05:01:50,348 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742286_1462 (size=8101) 2025-01-08T05:01:50,348 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742286_1462 (size=8101) 2025-01-08T05:01:50,348 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:01:50,353 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108d2bb5b0ab69241d7822a30d327032e54_150c2aaa34b578895b86a687c341c2e5 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/c4ca4238a0b923820dcc509a6f75849b20250108d2bb5b0ab69241d7822a30d327032e54_150c2aaa34b578895b86a687c341c2e5 2025-01-08T05:01:50,353 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/150c2aaa34b578895b86a687c341c2e5/.tmp/cf/ad7c5a2b5542411caaaf773584b152aa, store: [table=testtb-testExportWithChecksum family=cf region=150c2aaa34b578895b86a687c341c2e5] 2025-01-08T05:01:50,354 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/150c2aaa34b578895b86a687c341c2e5/.tmp/cf/ad7c5a2b5542411caaaf773584b152aa is 206, key is 19c544c9768b3220591f2e1ca68c6395c/cf:q/1736312510088/Put/seqid=0 2025-01-08T05:01:50,356 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742287_1463 (size=5172) 2025-01-08T05:01:50,356 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742287_1463 (size=5172) 2025-01-08T05:01:50,356 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742287_1463 (size=5172) 2025-01-08T05:01:50,357 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:01:50,359 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742288_1464 (size=14651) 2025-01-08T05:01:50,359 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742288_1464 (size=14651) 2025-01-08T05:01:50,360 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742288_1464 (size=14651) 2025-01-08T05:01:50,360 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=3.0 K, hasBloomFilter=true, into tmp file hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/150c2aaa34b578895b86a687c341c2e5/.tmp/cf/ad7c5a2b5542411caaaf773584b152aa 2025-01-08T05:01:50,361 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108cf3c5da2171f408b852005f2c74f7f44_0a15cae7e53ceb637248f515514255ec to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/d41d8cd98f00b204e9800998ecf8427e20250108cf3c5da2171f408b852005f2c74f7f44_0a15cae7e53ceb637248f515514255ec 2025-01-08T05:01:50,362 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/0a15cae7e53ceb637248f515514255ec/.tmp/cf/f3dd0a7de1cb4c0da2061e1487108377, store: [table=testtb-testExportWithChecksum family=cf region=0a15cae7e53ceb637248f515514255ec] 2025-01-08T05:01:50,363 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/0a15cae7e53ceb637248f515514255ec/.tmp/cf/f3dd0a7de1cb4c0da2061e1487108377 is 206, key is 019c138d40af3beeb17a62f0eff1ccf99/cf:q/1736312510087/Put/seqid=0 2025-01-08T05:01:50,366 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/150c2aaa34b578895b86a687c341c2e5/.tmp/cf/ad7c5a2b5542411caaaf773584b152aa as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/150c2aaa34b578895b86a687c341c2e5/cf/ad7c5a2b5542411caaaf773584b152aa 2025-01-08T05:01:50,370 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/150c2aaa34b578895b86a687c341c2e5/cf/ad7c5a2b5542411caaaf773584b152aa, entries=46, sequenceid=6, filesize=14.3 K 2025-01-08T05:01:50,371 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.HRegion(3040): Finished flush of dataSize ~3.00 KB/3070, heapSize ~6.70 KB/6864, currentSize=0 B/0 for 150c2aaa34b578895b86a687c341c2e5 in 60ms, sequenceid=6, compaction requested=false 2025-01-08T05:01:50,371 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportWithChecksum' 2025-01-08T05:01:50,372 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.HRegion(2538): Flush status journal for 150c2aaa34b578895b86a687c341c2e5: 2025-01-08T05:01:50,372 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5. for snaptb0-testExportWithChecksum completed. 2025-01-08T05:01:50,372 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5.' region-info for snapshot=snaptb0-testExportWithChecksum 2025-01-08T05:01:50,372 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T05:01:50,372 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/150c2aaa34b578895b86a687c341c2e5/cf/ad7c5a2b5542411caaaf773584b152aa] hfiles 2025-01-08T05:01:50,372 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/150c2aaa34b578895b86a687c341c2e5/cf/ad7c5a2b5542411caaaf773584b152aa for snapshot=snaptb0-testExportWithChecksum 2025-01-08T05:01:50,378 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742289_1465 (size=6108) 2025-01-08T05:01:50,378 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742289_1465 (size=6108) 2025-01-08T05:01:50,378 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742289_1465 (size=6108) 2025-01-08T05:01:50,379 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=266, hasBloomFilter=true, into tmp file hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/0a15cae7e53ceb637248f515514255ec/.tmp/cf/f3dd0a7de1cb4c0da2061e1487108377 2025-01-08T05:01:50,381 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742290_1466 (size=107) 2025-01-08T05:01:50,381 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742290_1466 (size=107) 2025-01-08T05:01:50,382 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742290_1466 (size=107) 2025-01-08T05:01:50,382 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5. 2025-01-08T05:01:50,382 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=190 2025-01-08T05:01:50,382 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=190 2025-01-08T05:01:50,383 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportWithChecksum on region 150c2aaa34b578895b86a687c341c2e5 2025-01-08T05:01:50,383 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=190, ppid=188, state=RUNNABLE; SnapshotRegionProcedure 150c2aaa34b578895b86a687c341c2e5 2025-01-08T05:01:50,383 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/0a15cae7e53ceb637248f515514255ec/.tmp/cf/f3dd0a7de1cb4c0da2061e1487108377 as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/0a15cae7e53ceb637248f515514255ec/cf/f3dd0a7de1cb4c0da2061e1487108377 2025-01-08T05:01:50,384 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=190, ppid=188, state=SUCCESS; SnapshotRegionProcedure 150c2aaa34b578895b86a687c341c2e5 in 225 msec 2025-01-08T05:01:50,390 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/0a15cae7e53ceb637248f515514255ec/cf/f3dd0a7de1cb4c0da2061e1487108377, entries=4, sequenceid=6, filesize=6.0 K 2025-01-08T05:01:50,391 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.HRegion(3040): Finished flush of dataSize ~266 B/266, heapSize ~816 B/816, currentSize=0 B/0 for 0a15cae7e53ceb637248f515514255ec in 80ms, sequenceid=6, compaction requested=false 2025-01-08T05:01:50,391 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.HRegion(2538): Flush status journal for 0a15cae7e53ceb637248f515514255ec: 2025-01-08T05:01:50,391 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec. for snaptb0-testExportWithChecksum completed. 2025-01-08T05:01:50,391 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec.' region-info for snapshot=snaptb0-testExportWithChecksum 2025-01-08T05:01:50,391 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T05:01:50,391 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/0a15cae7e53ceb637248f515514255ec/cf/f3dd0a7de1cb4c0da2061e1487108377] hfiles 2025-01-08T05:01:50,391 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/0a15cae7e53ceb637248f515514255ec/cf/f3dd0a7de1cb4c0da2061e1487108377 for snapshot=snaptb0-testExportWithChecksum 2025-01-08T05:01:50,396 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742291_1467 (size=107) 2025-01-08T05:01:50,397 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742291_1467 (size=107) 2025-01-08T05:01:50,397 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742291_1467 (size=107) 2025-01-08T05:01:50,397 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec. 2025-01-08T05:01:50,398 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=189 2025-01-08T05:01:50,398 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=189 2025-01-08T05:01:50,398 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportWithChecksum on region 0a15cae7e53ceb637248f515514255ec 2025-01-08T05:01:50,398 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=189, ppid=188, state=RUNNABLE; SnapshotRegionProcedure 0a15cae7e53ceb637248f515514255ec 2025-01-08T05:01:50,400 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=189, resume processing ppid=188 2025-01-08T05:01:50,400 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=189, ppid=188, state=SUCCESS; SnapshotRegionProcedure 0a15cae7e53ceb637248f515514255ec in 241 msec 2025-01-08T05:01:50,400 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T05:01:50,401 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T05:01:50,402 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T05:01:50,402 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T05:01:50,402 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:01:50,403 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/c4ca4238a0b923820dcc509a6f75849b20250108d2bb5b0ab69241d7822a30d327032e54_150c2aaa34b578895b86a687c341c2e5, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/d41d8cd98f00b204e9800998ecf8427e20250108cf3c5da2171f408b852005f2c74f7f44_0a15cae7e53ceb637248f515514255ec] hfiles 2025-01-08T05:01:50,403 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/c4ca4238a0b923820dcc509a6f75849b20250108d2bb5b0ab69241d7822a30d327032e54_150c2aaa34b578895b86a687c341c2e5 2025-01-08T05:01:50,403 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/d41d8cd98f00b204e9800998ecf8427e20250108cf3c5da2171f408b852005f2c74f7f44_0a15cae7e53ceb637248f515514255ec 2025-01-08T05:01:50,410 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742292_1468 (size=291) 2025-01-08T05:01:50,410 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742292_1468 (size=291) 2025-01-08T05:01:50,410 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742292_1468 (size=291) 2025-01-08T05:01:50,411 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T05:01:50,412 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testExportWithChecksum 2025-01-08T05:01:50,412 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/snaptb0-testExportWithChecksum 2025-01-08T05:01:50,433 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742293_1469 (size=951) 2025-01-08T05:01:50,433 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742293_1469 (size=951) 2025-01-08T05:01:50,434 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742293_1469 (size=951) 2025-01-08T05:01:50,440 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T05:01:50,440 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=188 2025-01-08T05:01:50,462 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T05:01:50,462 DEBUG [PEWorker-3 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/snaptb0-testExportWithChecksum to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportWithChecksum 2025-01-08T05:01:50,464 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T05:01:50,465 DEBUG [PEWorker-3 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 }, snapshot procedure id = 188 2025-01-08T05:01:50,466 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=188, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } in 328 msec 2025-01-08T05:01:50,742 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=188 2025-01-08T05:01:50,742 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportWithChecksum, procId: 188 completed 2025-01-08T05:01:50,742 INFO [Time-limited test {}] snapshot.TestExportSnapshot(476): Local export destination path: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312510742 2025-01-08T05:01:50,742 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=file:///, tgtDir=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312510742, rawTgtDir=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312510742, srcFsUri=hdfs://localhost:45939, srcDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:01:50,773 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:45939, inputRoot=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:01:50,773 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=org.apache.hadoop.fs.LocalFileSystem@2c860d76, outputRoot=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312510742, skipTmp=false, initialOutputSnapshotDir=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312510742/.hbase-snapshot/.tmp/snaptb0-testExportWithChecksum 2025-01-08T05:01:50,775 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T05:01:50,778 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportWithChecksum to file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312510742/.hbase-snapshot/.tmp/snaptb0-testExportWithChecksum 2025-01-08T05:01:50,801 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.7.0-SNAPSHOT.jar 2025-01-08T05:01:50,801 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T05:01:50,801 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T05:01:50,801 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T05:01:51,807 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop-7793345313009862572.jar 2025-01-08T05:01:51,807 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T05:01:51,808 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T05:01:51,880 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop-18159435152918305279.jar 2025-01-08T05:01:51,880 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T05:01:51,880 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T05:01:51,881 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T05:01:51,881 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T05:01:51,881 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T05:01:51,881 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T05:01:51,882 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T05:01:51,882 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T05:01:51,883 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T05:01:51,883 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T05:01:51,883 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T05:01:51,883 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T05:01:51,884 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T05:01:51,884 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T05:01:51,884 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T05:01:51,885 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T05:01:51,885 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T05:01:51,885 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T05:01:51,886 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:01:51,886 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:01:51,886 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T05:01:51,887 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:01:51,887 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:01:51,887 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T05:01:51,888 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T05:01:51,961 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742294_1470 (size=127628) 2025-01-08T05:01:51,961 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742294_1470 (size=127628) 2025-01-08T05:01:51,961 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742294_1470 (size=127628) 2025-01-08T05:01:51,989 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742295_1471 (size=2172137) 2025-01-08T05:01:51,989 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742295_1471 (size=2172137) 2025-01-08T05:01:51,989 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742295_1471 (size=2172137) 2025-01-08T05:01:51,997 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742296_1472 (size=213228) 2025-01-08T05:01:51,997 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742296_1472 (size=213228) 2025-01-08T05:01:51,998 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742296_1472 (size=213228) 2025-01-08T05:01:52,008 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742297_1473 (size=1877034) 2025-01-08T05:01:52,009 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742297_1473 (size=1877034) 2025-01-08T05:01:52,009 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742297_1473 (size=1877034) 2025-01-08T05:01:52,016 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742298_1474 (size=533455) 2025-01-08T05:01:52,016 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742298_1474 (size=533455) 2025-01-08T05:01:52,017 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742298_1474 (size=533455) 2025-01-08T05:01:52,038 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742299_1475 (size=7280644) 2025-01-08T05:01:52,038 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742299_1475 (size=7280644) 2025-01-08T05:01:52,039 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742299_1475 (size=7280644) 2025-01-08T05:01:52,057 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742300_1476 (size=4188619) 2025-01-08T05:01:52,058 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742300_1476 (size=4188619) 2025-01-08T05:01:52,058 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742300_1476 (size=4188619) 2025-01-08T05:01:52,066 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742301_1477 (size=20406) 2025-01-08T05:01:52,066 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742301_1477 (size=20406) 2025-01-08T05:01:52,067 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742301_1477 (size=20406) 2025-01-08T05:01:52,074 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742302_1478 (size=75495) 2025-01-08T05:01:52,074 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742302_1478 (size=75495) 2025-01-08T05:01:52,074 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742302_1478 (size=75495) 2025-01-08T05:01:52,080 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742303_1479 (size=45609) 2025-01-08T05:01:52,080 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742303_1479 (size=45609) 2025-01-08T05:01:52,081 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742303_1479 (size=45609) 2025-01-08T05:01:52,099 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742304_1480 (size=6350712) 2025-01-08T05:01:52,100 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742304_1480 (size=6350712) 2025-01-08T05:01:52,100 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742304_1480 (size=6350712) 2025-01-08T05:01:52,106 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742305_1481 (size=110084) 2025-01-08T05:01:52,107 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742305_1481 (size=110084) 2025-01-08T05:01:52,107 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742305_1481 (size=110084) 2025-01-08T05:01:52,116 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742306_1482 (size=1323991) 2025-01-08T05:01:52,116 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742306_1482 (size=1323991) 2025-01-08T05:01:52,116 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742306_1482 (size=1323991) 2025-01-08T05:01:52,122 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742307_1483 (size=23076) 2025-01-08T05:01:52,122 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742307_1483 (size=23076) 2025-01-08T05:01:52,122 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742307_1483 (size=23076) 2025-01-08T05:01:52,128 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742308_1484 (size=126803) 2025-01-08T05:01:52,128 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742308_1484 (size=126803) 2025-01-08T05:01:52,129 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742308_1484 (size=126803) 2025-01-08T05:01:52,135 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742309_1485 (size=322274) 2025-01-08T05:01:52,136 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742309_1485 (size=322274) 2025-01-08T05:01:52,136 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742309_1485 (size=322274) 2025-01-08T05:01:52,146 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742310_1486 (size=1832290) 2025-01-08T05:01:52,146 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742310_1486 (size=1832290) 2025-01-08T05:01:52,146 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742310_1486 (size=1832290) 2025-01-08T05:01:52,154 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742311_1487 (size=451756) 2025-01-08T05:01:52,154 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742311_1487 (size=451756) 2025-01-08T05:01:52,154 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742311_1487 (size=451756) 2025-01-08T05:01:52,178 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742312_1488 (size=30081) 2025-01-08T05:01:52,179 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742312_1488 (size=30081) 2025-01-08T05:01:52,179 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742312_1488 (size=30081) 2025-01-08T05:01:52,590 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742313_1489 (size=53616) 2025-01-08T05:01:52,591 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742313_1489 (size=53616) 2025-01-08T05:01:52,591 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742313_1489 (size=53616) 2025-01-08T05:01:52,600 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742314_1490 (size=29229) 2025-01-08T05:01:52,600 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742314_1490 (size=29229) 2025-01-08T05:01:52,601 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742314_1490 (size=29229) 2025-01-08T05:01:52,611 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742315_1491 (size=169089) 2025-01-08T05:01:52,611 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742315_1491 (size=169089) 2025-01-08T05:01:52,611 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742315_1491 (size=169089) 2025-01-08T05:01:52,631 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742316_1492 (size=5175431) 2025-01-08T05:01:52,631 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742316_1492 (size=5175431) 2025-01-08T05:01:52,632 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742316_1492 (size=5175431) 2025-01-08T05:01:52,638 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742317_1493 (size=136454) 2025-01-08T05:01:52,638 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742317_1493 (size=136454) 2025-01-08T05:01:52,639 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742317_1493 (size=136454) 2025-01-08T05:01:52,647 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742318_1494 (size=907470) 2025-01-08T05:01:52,647 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742318_1494 (size=907470) 2025-01-08T05:01:52,647 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742318_1494 (size=907470) 2025-01-08T05:01:52,660 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742319_1495 (size=3317408) 2025-01-08T05:01:52,660 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742319_1495 (size=3317408) 2025-01-08T05:01:52,661 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742319_1495 (size=3317408) 2025-01-08T05:01:52,667 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742320_1496 (size=503880) 2025-01-08T05:01:52,668 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742320_1496 (size=503880) 2025-01-08T05:01:52,668 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742320_1496 (size=503880) 2025-01-08T05:01:52,685 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742321_1497 (size=4695811) 2025-01-08T05:01:52,685 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742321_1497 (size=4695811) 2025-01-08T05:01:52,685 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742321_1497 (size=4695811) 2025-01-08T05:01:52,687 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T05:01:52,689 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'snaptb0-testExportWithChecksum' hfile list 2025-01-08T05:01:52,690 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(751): export split=0 size=33.2 K 2025-01-08T05:01:52,696 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742322_1498 (size=714) 2025-01-08T05:01:52,696 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742322_1498 (size=714) 2025-01-08T05:01:52,697 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742322_1498 (size=714) 2025-01-08T05:01:52,701 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742323_1499 (size=15) 2025-01-08T05:01:52,701 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742323_1499 (size=15) 2025-01-08T05:01:52,702 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742323_1499 (size=15) 2025-01-08T05:01:52,712 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742324_1500 (size=304927) 2025-01-08T05:01:52,712 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742324_1500 (size=304927) 2025-01-08T05:01:52,712 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742324_1500 (size=304927) 2025-01-08T05:01:52,903 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T05:01:52,903 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T05:01:52,905 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0007_000001 (auth:SIMPLE) from 127.0.0.1:45202 2025-01-08T05:01:52,916 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_0/usercache/jenkins/appcache/application_1736312298603_0007/container_1736312298603_0007_01_000001/launch_container.sh] 2025-01-08T05:01:52,916 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_0/usercache/jenkins/appcache/application_1736312298603_0007/container_1736312298603_0007_01_000001/container_tokens] 2025-01-08T05:01:52,916 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_0/usercache/jenkins/appcache/application_1736312298603_0007/container_1736312298603_0007_01_000001/sysfs] 2025-01-08T05:01:52,926 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0008_000001 (auth:SIMPLE) from 127.0.0.1:60474 2025-01-08T05:01:54,125 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T05:01:58,345 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportWithChecksum 2025-01-08T05:01:58,345 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportWithChecksum Metrics about Tables on a single HBase RegionServer 2025-01-08T05:01:58,346 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testEmptyExportFileSystemState 2025-01-08T05:01:58,357 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0008_000001 (auth:SIMPLE) from 127.0.0.1:40506 2025-01-08T05:01:58,627 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742325_1501 (size=350601) 2025-01-08T05:01:58,627 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742325_1501 (size=350601) 2025-01-08T05:01:58,628 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742325_1501 (size=350601) 2025-01-08T05:02:00,664 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0008_000001 (auth:SIMPLE) from 127.0.0.1:52774 2025-01-08T05:02:02,422 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 5881ce671bd325587aadcdae5691632b, had cached 0 bytes from a total of 13275 2025-01-08T05:02:02,422 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region bf5538cd1c1443d97a8c76cd68b482f8, had cached 0 bytes from a total of 7282 2025-01-08T05:02:03,849 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T05:02:05,188 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_3/usercache/jenkins/appcache/application_1736312298603_0008/container_1736312298603_0008_01_000002/launch_container.sh] 2025-01-08T05:02:05,188 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_3/usercache/jenkins/appcache/application_1736312298603_0008/container_1736312298603_0008_01_000002/container_tokens] 2025-01-08T05:02:05,188 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_3/usercache/jenkins/appcache/application_1736312298603_0008/container_1736312298603_0008_01_000002/sysfs] 2025-01-08T05:02:05,792 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. Error: java.io.IOException: Checksum mismatch between hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/150c2aaa34b578895b86a687c341c2e5/cf/ad7c5a2b5542411caaaf773584b152aa and file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312510742/archive/data/default/testtb-testExportWithChecksum/150c2aaa34b578895b86a687c341c2e5/cf/ad7c5a2b5542411caaaf773584b152aa. Input and output filesystems are of different types. Their checksum algorithms may be incompatible. You can choose file-level checksum validation via -Ddfs.checksum.combine.mode=COMPOSITE_CRC when block-sizes or filesystems are different. Or you can skip checksum-checks altogether with -no-checksum-verify, for the table backup scenario, you should use -i option to skip checksum-checks. (NOTE: By skipping checksums, one runs the risk of masking data-corruption during file-transfer.) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.verifyCopyResult(ExportSnapshot.java:596) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.copyFile(ExportSnapshot.java:332) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.map(ExportSnapshot.java:254) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.map(ExportSnapshot.java:180) at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:145) at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:800) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:348) at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:178) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:172) 2025-01-08T05:02:06,461 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0008_000001 (auth:SIMPLE) from 127.0.0.1:44658 Error: java.io.IOException: Checksum mismatch between hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/150c2aaa34b578895b86a687c341c2e5/cf/ad7c5a2b5542411caaaf773584b152aa and file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312510742/archive/data/default/testtb-testExportWithChecksum/150c2aaa34b578895b86a687c341c2e5/cf/ad7c5a2b5542411caaaf773584b152aa. Input and output filesystems are of different types. Their checksum algorithms may be incompatible. You can choose file-level checksum validation via -Ddfs.checksum.combine.mode=COMPOSITE_CRC when block-sizes or filesystems are different. Or you can skip checksum-checks altogether with -no-checksum-verify, for the table backup scenario, you should use -i option to skip checksum-checks. (NOTE: By skipping checksums, one runs the risk of masking data-corruption during file-transfer.) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.verifyCopyResult(ExportSnapshot.java:596) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.copyFile(ExportSnapshot.java:332) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.map(ExportSnapshot.java:254) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.map(ExportSnapshot.java:180) at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:145) at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:800) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:348) at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:178) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:172) 2025-01-08T05:02:11,483 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0008_000001 (auth:SIMPLE) from 127.0.0.1:44662 2025-01-08T05:02:15,142 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-1_1/usercache/jenkins/appcache/application_1736312298603_0008/container_1736312298603_0008_01_000003/launch_container.sh] 2025-01-08T05:02:15,150 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-1_1/usercache/jenkins/appcache/application_1736312298603_0008/container_1736312298603_0008_01_000003/container_tokens] 2025-01-08T05:02:15,151 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-1_1/usercache/jenkins/appcache/application_1736312298603_0008/container_1736312298603_0008_01_000003/sysfs] 2025-01-08T05:02:15,259 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-1_2/usercache/jenkins/appcache/application_1736312298603_0008/container_1736312298603_0008_01_000004/launch_container.sh] 2025-01-08T05:02:15,259 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-1_2/usercache/jenkins/appcache/application_1736312298603_0008/container_1736312298603_0008_01_000004/container_tokens] 2025-01-08T05:02:15,259 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-1_2/usercache/jenkins/appcache/application_1736312298603_0008/container_1736312298603_0008_01_000004/sysfs] 2025-01-08T05:02:15,563 DEBUG [master/f28094b72178:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 0a15cae7e53ceb637248f515514255ec changed from -1.0 to 0.0, refreshing cache 2025-01-08T05:02:15,563 DEBUG [master/f28094b72178:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region bf5538cd1c1443d97a8c76cd68b482f8 changed from -1.0 to 0.0, refreshing cache 2025-01-08T05:02:15,564 DEBUG [master/f28094b72178:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 150c2aaa34b578895b86a687c341c2e5 changed from -1.0 to 0.0, refreshing cache 2025-01-08T05:02:15,564 DEBUG [master/f28094b72178:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 5881ce671bd325587aadcdae5691632b changed from -1.0 to 0.0, refreshing cache Error: java.io.IOException: Checksum mismatch between hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/150c2aaa34b578895b86a687c341c2e5/cf/ad7c5a2b5542411caaaf773584b152aa and file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/local-export-1736312510742/archive/data/default/testtb-testExportWithChecksum/150c2aaa34b578895b86a687c341c2e5/cf/ad7c5a2b5542411caaaf773584b152aa. Input and output filesystems are of different types. Their checksum algorithms may be incompatible. You can choose file-level checksum validation via -Ddfs.checksum.combine.mode=COMPOSITE_CRC when block-sizes or filesystems are different. Or you can skip checksum-checks altogether with -no-checksum-verify, for the table backup scenario, you should use -i option to skip checksum-checks. (NOTE: By skipping checksums, one runs the risk of masking data-corruption during file-transfer.) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.verifyCopyResult(ExportSnapshot.java:596) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.copyFile(ExportSnapshot.java:332) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.map(ExportSnapshot.java:254) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.map(ExportSnapshot.java:180) at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:145) at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:800) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:348) at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:178) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:172) 2025-01-08T05:02:16,499 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0008_000001 (auth:SIMPLE) from 127.0.0.1:33168 2025-01-08T05:02:19,811 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742326_1502 (size=21340) 2025-01-08T05:02:19,812 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742326_1502 (size=21340) 2025-01-08T05:02:19,812 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742326_1502 (size=21340) 2025-01-08T05:02:19,830 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742327_1503 (size=460) 2025-01-08T05:02:19,830 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742327_1503 (size=460) 2025-01-08T05:02:19,831 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742327_1503 (size=460) 2025-01-08T05:02:19,901 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742328_1504 (size=21340) 2025-01-08T05:02:19,903 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742328_1504 (size=21340) 2025-01-08T05:02:19,903 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742328_1504 (size=21340) 2025-01-08T05:02:19,904 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_1/usercache/jenkins/appcache/application_1736312298603_0008/container_1736312298603_0008_01_000005/launch_container.sh] 2025-01-08T05:02:19,904 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_1/usercache/jenkins/appcache/application_1736312298603_0008/container_1736312298603_0008_01_000005/container_tokens] 2025-01-08T05:02:19,904 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_1/usercache/jenkins/appcache/application_1736312298603_0008/container_1736312298603_0008_01_000005/sysfs] 2025-01-08T05:02:19,971 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742329_1505 (size=350601) 2025-01-08T05:02:19,971 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742329_1505 (size=350601) 2025-01-08T05:02:19,971 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742329_1505 (size=350601) 2025-01-08T05:02:19,990 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0008_000001 (auth:SIMPLE) from 127.0.0.1:33182 2025-01-08T05:02:21,947 ERROR [Time-limited test {}] snapshot.ExportSnapshot(1227): Snapshot export failed org.apache.hadoop.hbase.snapshot.ExportSnapshotException: Task failed task_1736312298603_0008_m_000000 Job failed as tasks failed. failedMaps:1 failedReduces:0 killedMaps:0 killedReduces: 0 at org.apache.hadoop.hbase.snapshot.ExportSnapshot.runCopyJob(ExportSnapshot.java:935) ~[classes/:?] at org.apache.hadoop.hbase.snapshot.ExportSnapshot.doWork(ExportSnapshot.java:1204) ~[classes/:?] at org.apache.hadoop.hbase.util.AbstractHBaseTool.run(AbstractHBaseTool.java:151) ~[hbase-common-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:82) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hbase.snapshot.TestExportSnapshot.runExportSnapshot(TestExportSnapshot.java:523) ~[test-classes/:?] at org.apache.hadoop.hbase.snapshot.TestExportSnapshot.testExportFileSystemState(TestExportSnapshot.java:353) ~[test-classes/:?] at org.apache.hadoop.hbase.snapshot.TestExportSnapshot.testExportWithChecksum(TestExportSnapshot.java:237) ~[test-classes/:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) ~[junit-4.13.2.jar:4.13.2] at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) ~[junit-4.13.2.jar:4.13.2] at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2025-01-08T05:02:21,948 INFO [Time-limited test {}] snapshot.TestExportSnapshot(468): HDFS export destination path: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312541948 2025-01-08T05:02:21,948 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=hdfs://localhost:45939, tgtDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312541948, rawTgtDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312541948, srcFsUri=hdfs://localhost:45939, srcDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:02:21,984 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:45939, inputRoot=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:02:21,984 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1149249371_22, ugi=jenkins (auth:SIMPLE)]], outputRoot=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312541948, skipTmp=false, initialOutputSnapshotDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312541948/.hbase-snapshot/.tmp/snaptb0-testExportWithChecksum 2025-01-08T05:02:21,986 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T05:02:21,992 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportWithChecksum to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312541948/.hbase-snapshot/.tmp/snaptb0-testExportWithChecksum 2025-01-08T05:02:22,022 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742330_1506 (size=156) 2025-01-08T05:02:22,022 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742330_1506 (size=156) 2025-01-08T05:02:22,022 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742330_1506 (size=156) 2025-01-08T05:02:22,032 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742331_1507 (size=951) 2025-01-08T05:02:22,032 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742331_1507 (size=951) 2025-01-08T05:02:22,033 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742331_1507 (size=951) 2025-01-08T05:02:22,035 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.7.0-SNAPSHOT.jar 2025-01-08T05:02:22,035 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T05:02:22,036 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T05:02:22,036 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T05:02:23,316 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop-13630263477163885531.jar 2025-01-08T05:02:23,316 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T05:02:23,317 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T05:02:23,388 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop-3286955602834447834.jar 2025-01-08T05:02:23,388 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T05:02:23,389 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T05:02:23,389 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T05:02:23,389 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T05:02:23,390 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T05:02:23,390 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T05:02:23,390 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T05:02:23,390 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T05:02:23,390 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T05:02:23,391 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T05:02:23,391 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T05:02:23,391 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T05:02:23,391 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T05:02:23,391 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T05:02:23,391 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T05:02:23,391 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T05:02:23,392 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T05:02:23,392 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T05:02:23,392 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:02:23,392 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:02:23,392 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T05:02:23,393 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:02:23,393 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:02:23,393 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T05:02:23,393 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T05:02:23,446 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742332_1508 (size=127628) 2025-01-08T05:02:23,446 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742332_1508 (size=127628) 2025-01-08T05:02:23,446 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742332_1508 (size=127628) 2025-01-08T05:02:23,458 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742333_1509 (size=2172137) 2025-01-08T05:02:23,458 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742333_1509 (size=2172137) 2025-01-08T05:02:23,458 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742333_1509 (size=2172137) 2025-01-08T05:02:23,469 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742334_1510 (size=213228) 2025-01-08T05:02:23,470 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742334_1510 (size=213228) 2025-01-08T05:02:23,470 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742334_1510 (size=213228) 2025-01-08T05:02:23,493 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742335_1511 (size=1877034) 2025-01-08T05:02:23,494 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742335_1511 (size=1877034) 2025-01-08T05:02:23,494 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742335_1511 (size=1877034) 2025-01-08T05:02:23,516 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742336_1512 (size=533455) 2025-01-08T05:02:23,516 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742336_1512 (size=533455) 2025-01-08T05:02:23,516 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742336_1512 (size=533455) 2025-01-08T05:02:23,537 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742337_1513 (size=7280644) 2025-01-08T05:02:23,537 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742337_1513 (size=7280644) 2025-01-08T05:02:23,537 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742337_1513 (size=7280644) 2025-01-08T05:02:23,554 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742338_1514 (size=4188619) 2025-01-08T05:02:23,554 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742338_1514 (size=4188619) 2025-01-08T05:02:23,555 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742338_1514 (size=4188619) 2025-01-08T05:02:23,561 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742339_1515 (size=20406) 2025-01-08T05:02:23,561 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742339_1515 (size=20406) 2025-01-08T05:02:23,561 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742339_1515 (size=20406) 2025-01-08T05:02:23,567 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742340_1516 (size=75495) 2025-01-08T05:02:23,567 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742340_1516 (size=75495) 2025-01-08T05:02:23,567 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742340_1516 (size=75495) 2025-01-08T05:02:23,573 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742341_1517 (size=45609) 2025-01-08T05:02:23,573 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742341_1517 (size=45609) 2025-01-08T05:02:23,573 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742341_1517 (size=45609) 2025-01-08T05:02:23,583 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742342_1518 (size=110084) 2025-01-08T05:02:23,584 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742342_1518 (size=110084) 2025-01-08T05:02:23,584 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742342_1518 (size=110084) 2025-01-08T05:02:23,593 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742343_1519 (size=1323991) 2025-01-08T05:02:23,593 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742343_1519 (size=1323991) 2025-01-08T05:02:23,593 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742343_1519 (size=1323991) 2025-01-08T05:02:23,599 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742344_1520 (size=23076) 2025-01-08T05:02:23,599 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742344_1520 (size=23076) 2025-01-08T05:02:23,599 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742344_1520 (size=23076) 2025-01-08T05:02:23,605 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742345_1521 (size=126803) 2025-01-08T05:02:23,605 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742345_1521 (size=126803) 2025-01-08T05:02:23,605 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742345_1521 (size=126803) 2025-01-08T05:02:23,613 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742346_1522 (size=322274) 2025-01-08T05:02:23,613 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742346_1522 (size=322274) 2025-01-08T05:02:23,614 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742346_1522 (size=322274) 2025-01-08T05:02:23,632 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742347_1523 (size=1832290) 2025-01-08T05:02:23,638 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742347_1523 (size=1832290) 2025-01-08T05:02:23,638 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742347_1523 (size=1832290) 2025-01-08T05:02:23,653 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742348_1524 (size=30081) 2025-01-08T05:02:23,653 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742348_1524 (size=30081) 2025-01-08T05:02:23,654 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742348_1524 (size=30081) 2025-01-08T05:02:23,668 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742349_1525 (size=53616) 2025-01-08T05:02:23,668 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742349_1525 (size=53616) 2025-01-08T05:02:23,669 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742349_1525 (size=53616) 2025-01-08T05:02:23,687 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742350_1526 (size=29229) 2025-01-08T05:02:23,687 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742350_1526 (size=29229) 2025-01-08T05:02:23,688 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742350_1526 (size=29229) 2025-01-08T05:02:23,695 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742351_1527 (size=169089) 2025-01-08T05:02:23,695 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742351_1527 (size=169089) 2025-01-08T05:02:23,695 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742351_1527 (size=169089) 2025-01-08T05:02:23,702 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742352_1528 (size=451756) 2025-01-08T05:02:23,702 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742352_1528 (size=451756) 2025-01-08T05:02:23,702 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742352_1528 (size=451756) 2025-01-08T05:02:23,721 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742353_1529 (size=6350712) 2025-01-08T05:02:23,721 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742353_1529 (size=6350712) 2025-01-08T05:02:23,721 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742353_1529 (size=6350712) 2025-01-08T05:02:23,742 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742354_1530 (size=5175431) 2025-01-08T05:02:23,743 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742354_1530 (size=5175431) 2025-01-08T05:02:23,743 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742354_1530 (size=5175431) 2025-01-08T05:02:23,755 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742355_1531 (size=136454) 2025-01-08T05:02:23,755 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742355_1531 (size=136454) 2025-01-08T05:02:23,755 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742355_1531 (size=136454) 2025-01-08T05:02:23,764 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742356_1532 (size=907470) 2025-01-08T05:02:23,764 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742356_1532 (size=907470) 2025-01-08T05:02:23,765 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742356_1532 (size=907470) 2025-01-08T05:02:23,785 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742357_1533 (size=3317408) 2025-01-08T05:02:23,785 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742357_1533 (size=3317408) 2025-01-08T05:02:23,785 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742357_1533 (size=3317408) 2025-01-08T05:02:23,793 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742358_1534 (size=503880) 2025-01-08T05:02:23,793 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742358_1534 (size=503880) 2025-01-08T05:02:23,793 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742358_1534 (size=503880) 2025-01-08T05:02:23,812 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742359_1535 (size=4695811) 2025-01-08T05:02:23,813 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742359_1535 (size=4695811) 2025-01-08T05:02:23,813 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742359_1535 (size=4695811) 2025-01-08T05:02:23,813 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T05:02:23,816 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'snaptb0-testExportWithChecksum' hfile list 2025-01-08T05:02:23,817 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(751): export split=0 size=33.2 K 2025-01-08T05:02:23,822 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742360_1536 (size=714) 2025-01-08T05:02:23,823 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742360_1536 (size=714) 2025-01-08T05:02:23,823 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742360_1536 (size=714) 2025-01-08T05:02:23,827 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742361_1537 (size=15) 2025-01-08T05:02:23,827 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742361_1537 (size=15) 2025-01-08T05:02:23,828 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742361_1537 (size=15) 2025-01-08T05:02:23,837 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742362_1538 (size=304877) 2025-01-08T05:02:23,837 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742362_1538 (size=304877) 2025-01-08T05:02:23,838 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742362_1538 (size=304877) 2025-01-08T05:02:26,173 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T05:02:26,173 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T05:02:26,176 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0008_000001 (auth:SIMPLE) from 127.0.0.1:58878 2025-01-08T05:02:26,875 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0009_000001 (auth:SIMPLE) from 127.0.0.1:46330 2025-01-08T05:02:31,210 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-1_0/usercache/jenkins/appcache/application_1736312298603_0008/container_1736312298603_0008_01_000001/launch_container.sh] 2025-01-08T05:02:31,210 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-1_0/usercache/jenkins/appcache/application_1736312298603_0008/container_1736312298603_0008_01_000001/container_tokens] 2025-01-08T05:02:31,210 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-1_0/usercache/jenkins/appcache/application_1736312298603_0008/container_1736312298603_0008_01_000001/sysfs] 2025-01-08T05:02:32,494 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0009_000001 (auth:SIMPLE) from 127.0.0.1:60698 2025-01-08T05:02:32,727 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742363_1539 (size=350551) 2025-01-08T05:02:32,728 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742363_1539 (size=350551) 2025-01-08T05:02:32,728 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742363_1539 (size=350551) 2025-01-08T05:02:34,599 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 0a15cae7e53ceb637248f515514255ec, had cached 0 bytes from a total of 6108 2025-01-08T05:02:34,600 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 150c2aaa34b578895b86a687c341c2e5, had cached 0 bytes from a total of 14651 2025-01-08T05:02:34,779 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0009_000001 (auth:SIMPLE) from 127.0.0.1:55976 2025-01-08T05:02:35,792 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T05:02:37,918 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742364_1540 (size=14651) 2025-01-08T05:02:37,919 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742364_1540 (size=14651) 2025-01-08T05:02:37,919 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742364_1540 (size=14651) 2025-01-08T05:02:37,986 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742365_1541 (size=8101) 2025-01-08T05:02:37,986 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742365_1541 (size=8101) 2025-01-08T05:02:37,987 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742365_1541 (size=8101) 2025-01-08T05:02:38,041 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742366_1542 (size=6108) 2025-01-08T05:02:38,041 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742366_1542 (size=6108) 2025-01-08T05:02:38,041 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742366_1542 (size=6108) 2025-01-08T05:02:38,088 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742367_1543 (size=5172) 2025-01-08T05:02:38,088 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742367_1543 (size=5172) 2025-01-08T05:02:38,088 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742367_1543 (size=5172) 2025-01-08T05:02:38,168 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742368_1544 (size=17459) 2025-01-08T05:02:38,168 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742368_1544 (size=17459) 2025-01-08T05:02:38,169 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742368_1544 (size=17459) 2025-01-08T05:02:38,195 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742369_1545 (size=462) 2025-01-08T05:02:38,195 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742369_1545 (size=462) 2025-01-08T05:02:38,196 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742369_1545 (size=462) 2025-01-08T05:02:38,248 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_0/usercache/jenkins/appcache/application_1736312298603_0009/container_1736312298603_0009_01_000002/launch_container.sh] 2025-01-08T05:02:38,248 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_0/usercache/jenkins/appcache/application_1736312298603_0009/container_1736312298603_0009_01_000002/container_tokens] 2025-01-08T05:02:38,248 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_0/usercache/jenkins/appcache/application_1736312298603_0009/container_1736312298603_0009_01_000002/sysfs] 2025-01-08T05:02:38,256 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742370_1546 (size=17459) 2025-01-08T05:02:38,257 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742370_1546 (size=17459) 2025-01-08T05:02:38,257 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742370_1546 (size=17459) 2025-01-08T05:02:38,296 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742371_1547 (size=350551) 2025-01-08T05:02:38,296 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742371_1547 (size=350551) 2025-01-08T05:02:38,297 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742371_1547 (size=350551) 2025-01-08T05:02:38,310 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0009_000001 (auth:SIMPLE) from 127.0.0.1:55978 2025-01-08T05:02:39,994 INFO [Time-limited test {}] snapshot.ExportSnapshot(1207): Finalize the Snapshot Export 2025-01-08T05:02:39,995 INFO [Time-limited test {}] snapshot.ExportSnapshot(1218): Verify the exported snapshot's expiration status and integrity. 2025-01-08T05:02:40,008 INFO [Time-limited test {}] snapshot.ExportSnapshot(1224): Export Completed: snaptb0-testExportWithChecksum 2025-01-08T05:02:40,008 INFO [Time-limited test {}] snapshot.TestExportSnapshot(362): Exported snapshot 2025-01-08T05:02:40,009 INFO [Time-limited test {}] snapshot.TestExportSnapshot(373): Verified filesystem state 2025-01-08T05:02:40,009 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1149249371_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportWithChecksum at hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportWithChecksum 2025-01-08T05:02:40,009 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportWithChecksum/.snapshotinfo 2025-01-08T05:02:40,010 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportWithChecksum/data.manifest 2025-01-08T05:02:40,010 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1149249371_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312541948/.hbase-snapshot/snaptb0-testExportWithChecksum at hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312541948/.hbase-snapshot/snaptb0-testExportWithChecksum 2025-01-08T05:02:40,011 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312541948/.hbase-snapshot/snaptb0-testExportWithChecksum/.snapshotinfo 2025-01-08T05:02:40,011 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312541948/.hbase-snapshot/snaptb0-testExportWithChecksum/data.manifest 2025-01-08T05:02:40,023 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testExportWithChecksum 2025-01-08T05:02:40,023 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.3 disable testtb-testExportWithChecksum 2025-01-08T05:02:40,024 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=191, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testExportWithChecksum 2025-01-08T05:02:40,026 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=191 2025-01-08T05:02:40,027 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithChecksum","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312560026"}]},"ts":"1736312560026"} 2025-01-08T05:02:40,031 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithChecksum, state=DISABLING in hbase:meta 2025-01-08T05:02:40,033 INFO [PEWorker-1 {}] procedure.DisableTableProcedure(284): Set testtb-testExportWithChecksum to state=DISABLING 2025-01-08T05:02:40,034 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=192, ppid=191, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testExportWithChecksum}] 2025-01-08T05:02:40,036 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=193, ppid=192, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=0a15cae7e53ceb637248f515514255ec, UNASSIGN}, {pid=194, ppid=192, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=150c2aaa34b578895b86a687c341c2e5, UNASSIGN}] 2025-01-08T05:02:40,037 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=194, ppid=192, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=150c2aaa34b578895b86a687c341c2e5, UNASSIGN 2025-01-08T05:02:40,037 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=193, ppid=192, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=0a15cae7e53ceb637248f515514255ec, UNASSIGN 2025-01-08T05:02:40,038 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=194 updating hbase:meta row=150c2aaa34b578895b86a687c341c2e5, regionState=CLOSING, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T05:02:40,039 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=193 updating hbase:meta row=0a15cae7e53ceb637248f515514255ec, regionState=CLOSING, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T05:02:40,040 DEBUG [PEWorker-4 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T05:02:40,040 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=195, ppid=194, state=RUNNABLE; CloseRegionProcedure 150c2aaa34b578895b86a687c341c2e5, server=f28094b72178,37967,1736312288941}] 2025-01-08T05:02:40,042 DEBUG [PEWorker-5 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T05:02:40,042 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=196, ppid=193, state=RUNNABLE; CloseRegionProcedure 0a15cae7e53ceb637248f515514255ec, server=f28094b72178,37583,1736312288873}] 2025-01-08T05:02:40,127 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=191 2025-01-08T05:02:40,193 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:02:40,194 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] handler.UnassignRegionHandler(124): Close 150c2aaa34b578895b86a687c341c2e5 2025-01-08T05:02:40,194 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T05:02:40,194 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] regionserver.HRegion(1681): Closing 150c2aaa34b578895b86a687c341c2e5, disabling compactions & flushes 2025-01-08T05:02:40,194 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] regionserver.HRegion(1703): Closing region testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5. 2025-01-08T05:02:40,194 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5. 2025-01-08T05:02:40,194 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5. after waiting 0 ms 2025-01-08T05:02:40,194 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5. 2025-01-08T05:02:40,195 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T05:02:40,195 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] handler.UnassignRegionHandler(124): Close 0a15cae7e53ceb637248f515514255ec 2025-01-08T05:02:40,195 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T05:02:40,195 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] regionserver.HRegion(1681): Closing 0a15cae7e53ceb637248f515514255ec, disabling compactions & flushes 2025-01-08T05:02:40,195 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] regionserver.HRegion(1703): Closing region testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec. 2025-01-08T05:02:40,195 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec. 2025-01-08T05:02:40,195 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec. after waiting 0 ms 2025-01-08T05:02:40,195 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec. 2025-01-08T05:02:40,199 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/0a15cae7e53ceb637248f515514255ec/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T05:02:40,200 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/150c2aaa34b578895b86a687c341c2e5/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T05:02:40,200 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:02:40,200 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:02:40,200 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] regionserver.HRegion(1922): Closed testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec. 2025-01-08T05:02:40,200 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] regionserver.HRegion(1635): Region close journal for 0a15cae7e53ceb637248f515514255ec: 2025-01-08T05:02:40,200 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] regionserver.HRegion(1922): Closed testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5. 2025-01-08T05:02:40,200 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] regionserver.HRegion(1635): Region close journal for 150c2aaa34b578895b86a687c341c2e5: 2025-01-08T05:02:40,201 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] handler.UnassignRegionHandler(170): Closed 0a15cae7e53ceb637248f515514255ec 2025-01-08T05:02:40,202 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=193 updating hbase:meta row=0a15cae7e53ceb637248f515514255ec, regionState=CLOSED 2025-01-08T05:02:40,202 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] handler.UnassignRegionHandler(170): Closed 150c2aaa34b578895b86a687c341c2e5 2025-01-08T05:02:40,202 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=194 updating hbase:meta row=150c2aaa34b578895b86a687c341c2e5, regionState=CLOSED 2025-01-08T05:02:40,205 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=196, resume processing ppid=193 2025-01-08T05:02:40,206 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=195, resume processing ppid=194 2025-01-08T05:02:40,206 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=196, ppid=193, state=SUCCESS; CloseRegionProcedure 0a15cae7e53ceb637248f515514255ec, server=f28094b72178,37583,1736312288873 in 161 msec 2025-01-08T05:02:40,206 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=195, ppid=194, state=SUCCESS; CloseRegionProcedure 150c2aaa34b578895b86a687c341c2e5, server=f28094b72178,37967,1736312288941 in 164 msec 2025-01-08T05:02:40,206 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=193, ppid=192, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=0a15cae7e53ceb637248f515514255ec, UNASSIGN in 169 msec 2025-01-08T05:02:40,207 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=194, resume processing ppid=192 2025-01-08T05:02:40,207 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=194, ppid=192, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=150c2aaa34b578895b86a687c341c2e5, UNASSIGN in 170 msec 2025-01-08T05:02:40,209 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=192, resume processing ppid=191 2025-01-08T05:02:40,209 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=192, ppid=191, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testExportWithChecksum in 174 msec 2025-01-08T05:02:40,210 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithChecksum","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312560210"}]},"ts":"1736312560210"} 2025-01-08T05:02:40,211 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithChecksum, state=DISABLED in hbase:meta 2025-01-08T05:02:40,213 INFO [PEWorker-2 {}] procedure.DisableTableProcedure(296): Set testtb-testExportWithChecksum to state=DISABLED 2025-01-08T05:02:40,214 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=191, state=SUCCESS; DisableTableProcedure table=testtb-testExportWithChecksum in 190 msec 2025-01-08T05:02:40,328 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=191 2025-01-08T05:02:40,328 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testExportWithChecksum, procId: 191 completed 2025-01-08T05:02:40,329 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.3 delete testtb-testExportWithChecksum 2025-01-08T05:02:40,330 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=197, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testExportWithChecksum 2025-01-08T05:02:40,330 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=197, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportWithChecksum 2025-01-08T05:02:40,331 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testExportWithChecksum 2025-01-08T05:02:40,331 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=197, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testExportWithChecksum 2025-01-08T05:02:40,332 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37583 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testExportWithChecksum 2025-01-08T05:02:40,334 DEBUG [HFileArchiver-20 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/0a15cae7e53ceb637248f515514255ec 2025-01-08T05:02:40,334 DEBUG [HFileArchiver-21 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/150c2aaa34b578895b86a687c341c2e5 2025-01-08T05:02:40,336 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithChecksum 2025-01-08T05:02:40,336 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithChecksum 2025-01-08T05:02:40,336 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithChecksum 2025-01-08T05:02:40,336 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithChecksum with data PBUF 2025-01-08T05:02:40,336 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithChecksum with data PBUF 2025-01-08T05:02:40,338 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:02:40,338 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportWithChecksum 2025-01-08T05:02:40,338 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportWithChecksum 2025-01-08T05:02:40,338 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:02:40,338 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:02:40,338 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=197 2025-01-08T05:02:40,338 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithChecksum 2025-01-08T05:02:40,338 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:02:40,339 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithChecksum with data null 2025-01-08T05:02:40,339 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithChecksum with data null 2025-01-08T05:02:40,339 INFO [zk-permission-watcher-pool-0 {}] access.AuthManager(135): Skipping permission cache refresh because writable data is empty 2025-01-08T05:02:40,339 INFO [zk-permission-watcher-pool-0 {}] access.AuthManager(135): Skipping permission cache refresh because writable data is empty 2025-01-08T05:02:40,339 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:02:40,340 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:02:40,340 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:02:40,340 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:02:40,343 DEBUG [HFileArchiver-20 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/0a15cae7e53ceb637248f515514255ec/cf, FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/0a15cae7e53ceb637248f515514255ec/recovered.edits] 2025-01-08T05:02:40,343 DEBUG [HFileArchiver-21 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/150c2aaa34b578895b86a687c341c2e5/cf, FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/150c2aaa34b578895b86a687c341c2e5/recovered.edits] 2025-01-08T05:02:40,347 DEBUG [HFileArchiver-20 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/0a15cae7e53ceb637248f515514255ec/cf/f3dd0a7de1cb4c0da2061e1487108377 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportWithChecksum/0a15cae7e53ceb637248f515514255ec/cf/f3dd0a7de1cb4c0da2061e1487108377 2025-01-08T05:02:40,347 DEBUG [HFileArchiver-21 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/150c2aaa34b578895b86a687c341c2e5/cf/ad7c5a2b5542411caaaf773584b152aa to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportWithChecksum/150c2aaa34b578895b86a687c341c2e5/cf/ad7c5a2b5542411caaaf773584b152aa 2025-01-08T05:02:40,350 DEBUG [HFileArchiver-21 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/150c2aaa34b578895b86a687c341c2e5/recovered.edits/9.seqid to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportWithChecksum/150c2aaa34b578895b86a687c341c2e5/recovered.edits/9.seqid 2025-01-08T05:02:40,350 DEBUG [HFileArchiver-20 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/0a15cae7e53ceb637248f515514255ec/recovered.edits/9.seqid to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportWithChecksum/0a15cae7e53ceb637248f515514255ec/recovered.edits/9.seqid 2025-01-08T05:02:40,350 DEBUG [HFileArchiver-21 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/150c2aaa34b578895b86a687c341c2e5 2025-01-08T05:02:40,350 DEBUG [HFileArchiver-20 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportWithChecksum/0a15cae7e53ceb637248f515514255ec 2025-01-08T05:02:40,350 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(313): Archived testtb-testExportWithChecksum regions 2025-01-08T05:02:40,351 DEBUG [PEWorker-3 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6 2025-01-08T05:02:40,351 DEBUG [PEWorker-3 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf] 2025-01-08T05:02:40,354 DEBUG [PEWorker-3 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/c4ca4238a0b923820dcc509a6f75849b20250108d2bb5b0ab69241d7822a30d327032e54_150c2aaa34b578895b86a687c341c2e5 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/c4ca4238a0b923820dcc509a6f75849b20250108d2bb5b0ab69241d7822a30d327032e54_150c2aaa34b578895b86a687c341c2e5 2025-01-08T05:02:40,356 DEBUG [PEWorker-3 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/d41d8cd98f00b204e9800998ecf8427e20250108cf3c5da2171f408b852005f2c74f7f44_0a15cae7e53ceb637248f515514255ec to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/d41d8cd98f00b204e9800998ecf8427e20250108cf3c5da2171f408b852005f2c74f7f44_0a15cae7e53ceb637248f515514255ec 2025-01-08T05:02:40,356 DEBUG [PEWorker-3 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6 2025-01-08T05:02:40,358 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=197, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testExportWithChecksum 2025-01-08T05:02:40,360 WARN [PEWorker-3 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testtb-testExportWithChecksum from hbase:meta 2025-01-08T05:02:40,362 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testExportWithChecksum' descriptor. 2025-01-08T05:02:40,363 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=197, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testExportWithChecksum 2025-01-08T05:02:40,363 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testExportWithChecksum' from region states. 2025-01-08T05:02:40,364 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736312560363"}]},"ts":"9223372036854775807"} 2025-01-08T05:02:40,364 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736312560363"}]},"ts":"9223372036854775807"} 2025-01-08T05:02:40,366 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T05:02:40,366 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => 0a15cae7e53ceb637248f515514255ec, NAME => 'testtb-testExportWithChecksum,,1736312509133.0a15cae7e53ceb637248f515514255ec.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => 150c2aaa34b578895b86a687c341c2e5, NAME => 'testtb-testExportWithChecksum,1,1736312509133.150c2aaa34b578895b86a687c341c2e5.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T05:02:40,366 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testExportWithChecksum' as deleted. 2025-01-08T05:02:40,366 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportWithChecksum","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736312560366"}]},"ts":"9223372036854775807"} 2025-01-08T05:02:40,369 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testExportWithChecksum state from META 2025-01-08T05:02:40,375 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(133): Finished pid=197, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportWithChecksum 2025-01-08T05:02:40,379 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=197, state=SUCCESS; DeleteTableProcedure table=testtb-testExportWithChecksum in 46 msec 2025-01-08T05:02:40,439 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=197 2025-01-08T05:02:40,440 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testExportWithChecksum, procId: 197 completed 2025-01-08T05:02:40,448 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "emptySnaptb0-testExportWithChecksum" 2025-01-08T05:02:40,450 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotManager(380): Deleting snapshot: emptySnaptb0-testExportWithChecksum 2025-01-08T05:02:40,453 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "snaptb0-testExportWithChecksum" 2025-01-08T05:02:40,455 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testExportWithChecksum 2025-01-08T05:02:40,485 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: snapshot.TestMobSecureExportSnapshot#testExportWithChecksum Thread=802 (was 803), OpenFileDescriptor=807 (was 806) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=721 (was 931), ProcessCount=16 (was 16), AvailableMemoryMB=533 (was 1373) 2025-01-08T05:02:40,485 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=802 is superior to 500 2025-01-08T05:02:40,507 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: snapshot.TestMobSecureExportSnapshot#testExportFileSystemStateWithSkipTmp Thread=802, OpenFileDescriptor=807, MaxFileDescriptor=1048576, SystemLoadAverage=721, ProcessCount=16, AvailableMemoryMB=531 2025-01-08T05:02:40,507 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=802 is superior to 500 2025-01-08T05:02:40,513 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.3 create 'testtb-testExportFileSystemStateWithSkipTmp', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T05:02:40,514 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=198, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T05:02:40,516 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=198, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T05:02:40,516 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "testtb-testExportFileSystemStateWithSkipTmp" procId is: 198 2025-01-08T05:02:40,518 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=198 2025-01-08T05:02:40,519 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=198, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T05:02:40,535 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742372_1548 (size=454) 2025-01-08T05:02:40,535 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742372_1548 (size=454) 2025-01-08T05:02:40,535 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742372_1548 (size=454) 2025-01-08T05:02:40,619 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=198 2025-01-08T05:02:40,819 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=198 2025-01-08T05:02:40,934 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => ce04d37ae25166e295c48b46d4645627, NAME => 'testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testtb-testExportFileSystemStateWithSkipTmp', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:02:40,934 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => 6a4836669ea34b387dc723203ce2f74c, NAME => 'testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testtb-testExportFileSystemStateWithSkipTmp', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:02:40,940 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742374_1550 (size=79) 2025-01-08T05:02:40,941 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742373_1549 (size=79) 2025-01-08T05:02:40,941 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742374_1550 (size=79) 2025-01-08T05:02:40,941 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742374_1550 (size=79) 2025-01-08T05:02:40,941 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742373_1549 (size=79) 2025-01-08T05:02:40,941 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:02:40,941 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742373_1549 (size=79) 2025-01-08T05:02:40,941 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-0 {}] regionserver.HRegion(1681): Closing ce04d37ae25166e295c48b46d4645627, disabling compactions & flushes 2025-01-08T05:02:40,941 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:02:40,941 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627. 2025-01-08T05:02:40,941 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627. 2025-01-08T05:02:40,941 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627. after waiting 0 ms 2025-01-08T05:02:40,941 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627. 2025-01-08T05:02:40,941 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627. 2025-01-08T05:02:40,941 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-1 {}] regionserver.HRegion(1681): Closing 6a4836669ea34b387dc723203ce2f74c, disabling compactions & flushes 2025-01-08T05:02:40,941 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-0 {}] regionserver.HRegion(1635): Region close journal for ce04d37ae25166e295c48b46d4645627: 2025-01-08T05:02:40,941 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c. 2025-01-08T05:02:40,942 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c. 2025-01-08T05:02:40,942 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c. after waiting 0 ms 2025-01-08T05:02:40,942 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c. 2025-01-08T05:02:40,942 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c. 2025-01-08T05:02:40,942 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-1 {}] regionserver.HRegion(1635): Region close journal for 6a4836669ea34b387dc723203ce2f74c: 2025-01-08T05:02:40,942 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=198, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T05:02:40,943 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627.","families":{"info":[{"qualifier":"regioninfo","vlen":78,"tag":[],"timestamp":"1736312560943"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736312560943"}]},"ts":"1736312560943"} 2025-01-08T05:02:40,943 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c.","families":{"info":[{"qualifier":"regioninfo","vlen":78,"tag":[],"timestamp":"1736312560943"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736312560943"}]},"ts":"1736312560943"} 2025-01-08T05:02:40,945 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T05:02:40,945 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=198, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T05:02:40,945 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithSkipTmp","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312560945"}]},"ts":"1736312560945"} 2025-01-08T05:02:40,946 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithSkipTmp, state=ENABLING in hbase:meta 2025-01-08T05:02:40,950 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(202): Hosts are {f28094b72178=0} racks are {/default-rack=0} 2025-01-08T05:02:40,951 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T05:02:40,951 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T05:02:40,951 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T05:02:40,951 INFO [PEWorker-5 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T05:02:40,951 INFO [PEWorker-5 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T05:02:40,951 INFO [PEWorker-5 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T05:02:40,951 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T05:02:40,951 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=199, ppid=198, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=ce04d37ae25166e295c48b46d4645627, ASSIGN}, {pid=200, ppid=198, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=6a4836669ea34b387dc723203ce2f74c, ASSIGN}] 2025-01-08T05:02:40,952 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=200, ppid=198, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=6a4836669ea34b387dc723203ce2f74c, ASSIGN 2025-01-08T05:02:40,952 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=199, ppid=198, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=ce04d37ae25166e295c48b46d4645627, ASSIGN 2025-01-08T05:02:40,952 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=200, ppid=198, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=6a4836669ea34b387dc723203ce2f74c, ASSIGN; state=OFFLINE, location=f28094b72178,37583,1736312288873; forceNewPlan=false, retain=false 2025-01-08T05:02:40,952 INFO [PEWorker-1 {}] assignment.TransitRegionStateProcedure(264): Starting pid=199, ppid=198, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=ce04d37ae25166e295c48b46d4645627, ASSIGN; state=OFFLINE, location=f28094b72178,37967,1736312288941; forceNewPlan=false, retain=false 2025-01-08T05:02:41,103 INFO [f28094b72178:42195 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T05:02:41,103 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=200 updating hbase:meta row=6a4836669ea34b387dc723203ce2f74c, regionState=OPENING, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T05:02:41,103 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=199 updating hbase:meta row=ce04d37ae25166e295c48b46d4645627, regionState=OPENING, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T05:02:41,104 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=201, ppid=200, state=RUNNABLE; OpenRegionProcedure 6a4836669ea34b387dc723203ce2f74c, server=f28094b72178,37583,1736312288873}] 2025-01-08T05:02:41,105 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=202, ppid=199, state=RUNNABLE; OpenRegionProcedure ce04d37ae25166e295c48b46d4645627, server=f28094b72178,37967,1736312288941}] 2025-01-08T05:02:41,120 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=198 2025-01-08T05:02:41,256 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T05:02:41,256 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:02:41,259 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] handler.AssignRegionHandler(135): Open testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c. 2025-01-08T05:02:41,259 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] handler.AssignRegionHandler(135): Open testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627. 2025-01-08T05:02:41,259 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(7285): Opening region: {ENCODED => ce04d37ae25166e295c48b46d4645627, NAME => 'testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T05:02:41,259 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(7285): Opening region: {ENCODED => 6a4836669ea34b387dc723203ce2f74c, NAME => 'testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T05:02:41,259 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627. service=AccessControlService 2025-01-08T05:02:41,259 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c. service=AccessControlService 2025-01-08T05:02:41,259 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T05:02:41,259 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T05:02:41,259 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportFileSystemStateWithSkipTmp ce04d37ae25166e295c48b46d4645627 2025-01-08T05:02:41,259 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportFileSystemStateWithSkipTmp 6a4836669ea34b387dc723203ce2f74c 2025-01-08T05:02:41,260 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:02:41,260 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:02:41,260 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(7327): checking encryption for ce04d37ae25166e295c48b46d4645627 2025-01-08T05:02:41,260 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(7327): checking encryption for 6a4836669ea34b387dc723203ce2f74c 2025-01-08T05:02:41,260 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(7330): checking classloading for ce04d37ae25166e295c48b46d4645627 2025-01-08T05:02:41,260 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(7330): checking classloading for 6a4836669ea34b387dc723203ce2f74c 2025-01-08T05:02:41,261 INFO [StoreOpener-6a4836669ea34b387dc723203ce2f74c-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 6a4836669ea34b387dc723203ce2f74c 2025-01-08T05:02:41,261 INFO [StoreOpener-ce04d37ae25166e295c48b46d4645627-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region ce04d37ae25166e295c48b46d4645627 2025-01-08T05:02:41,262 INFO [StoreOpener-6a4836669ea34b387dc723203ce2f74c-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 6a4836669ea34b387dc723203ce2f74c columnFamilyName cf 2025-01-08T05:02:41,262 INFO [StoreOpener-ce04d37ae25166e295c48b46d4645627-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region ce04d37ae25166e295c48b46d4645627 columnFamilyName cf 2025-01-08T05:02:41,263 DEBUG [StoreOpener-ce04d37ae25166e295c48b46d4645627-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:02:41,263 DEBUG [StoreOpener-6a4836669ea34b387dc723203ce2f74c-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:02:41,263 INFO [StoreOpener-ce04d37ae25166e295c48b46d4645627-1 {}] regionserver.HStore(327): Store=ce04d37ae25166e295c48b46d4645627/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T05:02:41,263 INFO [StoreOpener-6a4836669ea34b387dc723203ce2f74c-1 {}] regionserver.HStore(327): Store=6a4836669ea34b387dc723203ce2f74c/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T05:02:41,264 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/ce04d37ae25166e295c48b46d4645627 2025-01-08T05:02:41,264 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/6a4836669ea34b387dc723203ce2f74c 2025-01-08T05:02:41,264 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/ce04d37ae25166e295c48b46d4645627 2025-01-08T05:02:41,264 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/6a4836669ea34b387dc723203ce2f74c 2025-01-08T05:02:41,266 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(1085): writing seq id for ce04d37ae25166e295c48b46d4645627 2025-01-08T05:02:41,266 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(1085): writing seq id for 6a4836669ea34b387dc723203ce2f74c 2025-01-08T05:02:41,267 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/ce04d37ae25166e295c48b46d4645627/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T05:02:41,267 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/6a4836669ea34b387dc723203ce2f74c/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T05:02:41,268 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(1102): Opened ce04d37ae25166e295c48b46d4645627; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=69552308, jitterRate=0.03641015291213989}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T05:02:41,268 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(1102): Opened 6a4836669ea34b387dc723203ce2f74c; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=75484805, jitterRate=0.12481124699115753}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T05:02:41,268 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(1001): Region open journal for ce04d37ae25166e295c48b46d4645627: 2025-01-08T05:02:41,268 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(1001): Region open journal for 6a4836669ea34b387dc723203ce2f74c: 2025-01-08T05:02:41,269 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627., pid=202, masterSystemTime=1736312561256 2025-01-08T05:02:41,269 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c., pid=201, masterSystemTime=1736312561256 2025-01-08T05:02:41,270 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627. 2025-01-08T05:02:41,270 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] handler.AssignRegionHandler(164): Opened testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627. 2025-01-08T05:02:41,270 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=199 updating hbase:meta row=ce04d37ae25166e295c48b46d4645627, regionState=OPEN, openSeqNum=2, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T05:02:41,270 DEBUG [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c. 2025-01-08T05:02:41,270 INFO [RS_OPEN_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] handler.AssignRegionHandler(164): Opened testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c. 2025-01-08T05:02:41,271 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=200 updating hbase:meta row=6a4836669ea34b387dc723203ce2f74c, regionState=OPEN, openSeqNum=2, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T05:02:41,273 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=202, resume processing ppid=199 2025-01-08T05:02:41,274 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=201, resume processing ppid=200 2025-01-08T05:02:41,274 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=201, ppid=200, state=SUCCESS; OpenRegionProcedure 6a4836669ea34b387dc723203ce2f74c, server=f28094b72178,37583,1736312288873 in 168 msec 2025-01-08T05:02:41,274 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=202, ppid=199, state=SUCCESS; OpenRegionProcedure ce04d37ae25166e295c48b46d4645627, server=f28094b72178,37967,1736312288941 in 166 msec 2025-01-08T05:02:41,274 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=199, ppid=198, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=ce04d37ae25166e295c48b46d4645627, ASSIGN in 322 msec 2025-01-08T05:02:41,275 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=200, resume processing ppid=198 2025-01-08T05:02:41,275 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=200, ppid=198, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=6a4836669ea34b387dc723203ce2f74c, ASSIGN in 323 msec 2025-01-08T05:02:41,276 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=198, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T05:02:41,276 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithSkipTmp","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312561276"}]},"ts":"1736312561276"} 2025-01-08T05:02:41,277 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithSkipTmp, state=ENABLED in hbase:meta 2025-01-08T05:02:41,286 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=198, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T05:02:41,287 DEBUG [PEWorker-1 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testExportFileSystemStateWithSkipTmp jenkins: RWXCA 2025-01-08T05:02:41,288 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37583 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemStateWithSkipTmp], kv [jenkins: RWXCA] 2025-01-08T05:02:41,298 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:02:41,298 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:02:41,298 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:02:41,298 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:02:41,300 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:02:41,301 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:02:41,301 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithSkipTmp with data PBUF\x0AQ\x0A\x07jenkins\x12F\x08\x03"B\x0A6\x0A\x07default\x12+testtb-testExportFileSystemStateWithSkipTmp \x00 \x01 \x02 \x03 \x04 2025-01-08T05:02:41,301 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithSkipTmp with data PBUF\x0AQ\x0A\x07jenkins\x12F\x08\x03"B\x0A6\x0A\x07default\x12+testtb-testExportFileSystemStateWithSkipTmp \x00 \x01 \x02 \x03 \x04 2025-01-08T05:02:41,301 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:02:41,301 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithSkipTmp with data PBUF\x0AQ\x0A\x07jenkins\x12F\x08\x03"B\x0A6\x0A\x07default\x12+testtb-testExportFileSystemStateWithSkipTmp \x00 \x01 \x02 \x03 \x04 2025-01-08T05:02:41,301 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:02:41,302 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithSkipTmp with data PBUF\x0AQ\x0A\x07jenkins\x12F\x08\x03"B\x0A6\x0A\x07default\x12+testtb-testExportFileSystemStateWithSkipTmp \x00 \x01 \x02 \x03 \x04 2025-01-08T05:02:41,303 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=198, state=SUCCESS; CreateTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp in 787 msec 2025-01-08T05:02:41,621 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=198 2025-01-08T05:02:41,621 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testExportFileSystemStateWithSkipTmp, procId: 198 completed 2025-01-08T05:02:41,624 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T05:02:41,624 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627. 2025-01-08T05:02:41,624 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T05:02:41,643 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } 2025-01-08T05:02:41,643 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736312561643 (current time:1736312561643). 2025-01-08T05:02:41,643 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T05:02:41,643 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot emptySnaptb0-testExportFileSystemStateWithSkipTmp VERSION not specified, setting to 2 2025-01-08T05:02:41,643 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T05:02:41,644 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x62e9c999 to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@7f116f19 2025-01-08T05:02:41,651 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@71b6a9e7, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:02:41,654 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:02:41,655 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:54308, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:02:41,656 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x62e9c999 to 127.0.0.1:59048 2025-01-08T05:02:41,656 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:02:41,657 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x7e474cec to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@a5a9d9a 2025-01-08T05:02:41,661 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4e67e4a7, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:02:41,663 DEBUG [hconnection-0x6b1db85c-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:02:41,664 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:54316, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:02:41,666 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x7e474cec to 127.0.0.1:59048 2025-01-08T05:02:41,666 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:02:41,666 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemStateWithSkipTmp], kv [jenkins: RWXCA] 2025-01-08T05:02:41,667 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T05:02:41,668 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=203, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } 2025-01-08T05:02:41,668 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 }, snapshot procedure id = 203 2025-01-08T05:02:41,669 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=203 2025-01-08T05:02:41,670 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T05:02:41,671 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T05:02:41,674 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T05:02:41,693 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742375_1551 (size=203) 2025-01-08T05:02:41,693 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742375_1551 (size=203) 2025-01-08T05:02:41,693 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742375_1551 (size=203) 2025-01-08T05:02:41,695 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T05:02:41,695 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=204, ppid=203, state=RUNNABLE; SnapshotRegionProcedure ce04d37ae25166e295c48b46d4645627}, {pid=205, ppid=203, state=RUNNABLE; SnapshotRegionProcedure 6a4836669ea34b387dc723203ce2f74c}] 2025-01-08T05:02:41,696 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=205, ppid=203, state=RUNNABLE; SnapshotRegionProcedure 6a4836669ea34b387dc723203ce2f74c 2025-01-08T05:02:41,696 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=204, ppid=203, state=RUNNABLE; SnapshotRegionProcedure ce04d37ae25166e295c48b46d4645627 2025-01-08T05:02:41,770 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=203 2025-01-08T05:02:41,847 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T05:02:41,847 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:02:41,848 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=37967 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=204 2025-01-08T05:02:41,848 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=37583 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=205 2025-01-08T05:02:41,848 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=204}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627. 2025-01-08T05:02:41,848 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=205}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c. 2025-01-08T05:02:41,848 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=204}] regionserver.HRegion(2538): Flush status journal for ce04d37ae25166e295c48b46d4645627: 2025-01-08T05:02:41,848 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=205}] regionserver.HRegion(2538): Flush status journal for 6a4836669ea34b387dc723203ce2f74c: 2025-01-08T05:02:41,848 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=204}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627. for emptySnaptb0-testExportFileSystemStateWithSkipTmp completed. 2025-01-08T05:02:41,848 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=205}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c. for emptySnaptb0-testExportFileSystemStateWithSkipTmp completed. 2025-01-08T05:02:41,848 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=204}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627.' region-info for snapshot=emptySnaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T05:02:41,848 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=205}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c.' region-info for snapshot=emptySnaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T05:02:41,848 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=204}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T05:02:41,848 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=205}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T05:02:41,848 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=204}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T05:02:41,848 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=205}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T05:02:41,854 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742376_1552 (size=82) 2025-01-08T05:02:41,854 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742377_1553 (size=82) 2025-01-08T05:02:41,854 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742376_1552 (size=82) 2025-01-08T05:02:41,854 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742376_1552 (size=82) 2025-01-08T05:02:41,855 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742377_1553 (size=82) 2025-01-08T05:02:41,855 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=204}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627. 2025-01-08T05:02:41,855 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=204}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=204 2025-01-08T05:02:41,855 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742377_1553 (size=82) 2025-01-08T05:02:41,856 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=205}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c. 2025-01-08T05:02:41,856 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=205}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=205 2025-01-08T05:02:41,856 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=204 2025-01-08T05:02:41,856 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportFileSystemStateWithSkipTmp on region ce04d37ae25166e295c48b46d4645627 2025-01-08T05:02:41,856 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=205 2025-01-08T05:02:41,856 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportFileSystemStateWithSkipTmp on region 6a4836669ea34b387dc723203ce2f74c 2025-01-08T05:02:41,856 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=204, ppid=203, state=RUNNABLE; SnapshotRegionProcedure ce04d37ae25166e295c48b46d4645627 2025-01-08T05:02:41,856 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=205, ppid=203, state=RUNNABLE; SnapshotRegionProcedure 6a4836669ea34b387dc723203ce2f74c 2025-01-08T05:02:41,857 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=204, ppid=203, state=SUCCESS; SnapshotRegionProcedure ce04d37ae25166e295c48b46d4645627 in 162 msec 2025-01-08T05:02:41,858 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=205, resume processing ppid=203 2025-01-08T05:02:41,858 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=205, ppid=203, state=SUCCESS; SnapshotRegionProcedure 6a4836669ea34b387dc723203ce2f74c in 162 msec 2025-01-08T05:02:41,858 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T05:02:41,859 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T05:02:41,860 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T05:02:41,860 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T05:02:41,860 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:02:41,860 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(328): No files under family: cf 2025-01-08T05:02:41,865 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742378_1554 (size=74) 2025-01-08T05:02:41,865 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742378_1554 (size=74) 2025-01-08T05:02:41,865 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742378_1554 (size=74) 2025-01-08T05:02:41,866 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T05:02:41,866 DEBUG [PEWorker-2 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for emptySnaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T05:02:41,866 DEBUG [PEWorker-2 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/emptySnaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T05:02:41,872 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742379_1555 (size=697) 2025-01-08T05:02:41,872 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742379_1555 (size=697) 2025-01-08T05:02:41,872 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742379_1555 (size=697) 2025-01-08T05:02:41,874 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T05:02:41,878 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T05:02:41,878 DEBUG [PEWorker-2 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/emptySnaptb0-testExportFileSystemStateWithSkipTmp to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/emptySnaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T05:02:41,879 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T05:02:41,879 DEBUG [PEWorker-2 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 }, snapshot procedure id = 203 2025-01-08T05:02:41,880 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=203, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } in 212 msec 2025-01-08T05:02:41,971 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=203 2025-01-08T05:02:41,972 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportFileSystemStateWithSkipTmp, procId: 203 completed 2025-01-08T05:02:41,975 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37967 {}] regionserver.HRegion(8254): writing data to region testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T05:02:41,976 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37583 {}] regionserver.HRegion(8254): writing data to region testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T05:02:41,979 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T05:02:41,979 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627. 2025-01-08T05:02:41,979 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T05:02:41,996 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } 2025-01-08T05:02:41,996 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736312561996 (current time:1736312561996). 2025-01-08T05:02:41,996 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T05:02:41,996 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testExportFileSystemStateWithSkipTmp VERSION not specified, setting to 2 2025-01-08T05:02:41,996 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T05:02:41,997 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x0022d78e to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@67b408e 2025-01-08T05:02:42,002 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@642f37ff, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:02:42,004 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:02:42,006 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:54322, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:02:42,007 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x0022d78e to 127.0.0.1:59048 2025-01-08T05:02:42,007 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:02:42,008 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x09f34194 to 127.0.0.1:59048 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@55501f14 2025-01-08T05:02:42,011 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@48dd4a8e, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T05:02:42,013 DEBUG [hconnection-0xc3809bc-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:02:42,014 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:54334, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:02:42,017 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x09f34194 to 127.0.0.1:59048 2025-01-08T05:02:42,017 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:02:42,017 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemStateWithSkipTmp], kv [jenkins: RWXCA] 2025-01-08T05:02:42,018 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T05:02:42,019 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=206, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } 2025-01-08T05:02:42,019 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 }, snapshot procedure id = 206 2025-01-08T05:02:42,020 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T05:02:42,020 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=206 2025-01-08T05:02:42,021 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T05:02:42,024 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T05:02:42,029 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742380_1556 (size=198) 2025-01-08T05:02:42,029 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742380_1556 (size=198) 2025-01-08T05:02:42,029 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742380_1556 (size=198) 2025-01-08T05:02:42,030 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T05:02:42,030 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=207, ppid=206, state=RUNNABLE; SnapshotRegionProcedure ce04d37ae25166e295c48b46d4645627}, {pid=208, ppid=206, state=RUNNABLE; SnapshotRegionProcedure 6a4836669ea34b387dc723203ce2f74c}] 2025-01-08T05:02:42,031 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=208, ppid=206, state=RUNNABLE; SnapshotRegionProcedure 6a4836669ea34b387dc723203ce2f74c 2025-01-08T05:02:42,031 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=207, ppid=206, state=RUNNABLE; SnapshotRegionProcedure ce04d37ae25166e295c48b46d4645627 2025-01-08T05:02:42,121 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=206 2025-01-08T05:02:42,156 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportFileSystemStateWithSkipTmp' 2025-01-08T05:02:42,182 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T05:02:42,182 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:02:42,183 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=37583 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=208 2025-01-08T05:02:42,183 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37967 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=207 2025-01-08T05:02:42,183 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c. 2025-01-08T05:02:42,183 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627. 2025-01-08T05:02:42,183 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.HRegion(2837): Flushing ce04d37ae25166e295c48b46d4645627 1/1 column families, dataSize=65 B heapSize=400 B 2025-01-08T05:02:42,183 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.HRegion(2837): Flushing 6a4836669ea34b387dc723203ce2f74c 1/1 column families, dataSize=3.19 KB heapSize=7.14 KB 2025-01-08T05:02:42,206 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b202501089a38c686adc2416d83a16cba30081c05_6a4836669ea34b387dc723203ce2f74c is 71, key is 1574a87a947f329d3f16033bc5bcace8/cf:q/1736312561976/Put/seqid=0 2025-01-08T05:02:42,208 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e202501088c51d0a25a30410fbe457a2e0b1c281c_ce04d37ae25166e295c48b46d4645627 is 69, key is 0f7c331fb75e1648f67459b1d3b9f1da3/cf:q/1736312561974/Put/seqid=0 2025-01-08T05:02:42,219 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742381_1557 (size=4964) 2025-01-08T05:02:42,219 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742381_1557 (size=4964) 2025-01-08T05:02:42,219 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742381_1557 (size=4964) 2025-01-08T05:02:42,219 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742382_1558 (size=8311) 2025-01-08T05:02:42,219 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742382_1558 (size=8311) 2025-01-08T05:02:42,219 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742382_1558 (size=8311) 2025-01-08T05:02:42,322 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=206 2025-01-08T05:02:42,620 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:02:42,620 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:02:42,624 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=206 2025-01-08T05:02:42,624 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b202501089a38c686adc2416d83a16cba30081c05_6a4836669ea34b387dc723203ce2f74c to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/c4ca4238a0b923820dcc509a6f75849b202501089a38c686adc2416d83a16cba30081c05_6a4836669ea34b387dc723203ce2f74c 2025-01-08T05:02:42,624 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e202501088c51d0a25a30410fbe457a2e0b1c281c_ce04d37ae25166e295c48b46d4645627 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/d41d8cd98f00b204e9800998ecf8427e202501088c51d0a25a30410fbe457a2e0b1c281c_ce04d37ae25166e295c48b46d4645627 2025-01-08T05:02:42,625 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/6a4836669ea34b387dc723203ce2f74c/.tmp/cf/090def20f49d4f809919df86a1b6a95b, store: [table=testtb-testExportFileSystemStateWithSkipTmp family=cf region=6a4836669ea34b387dc723203ce2f74c] 2025-01-08T05:02:42,626 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/ce04d37ae25166e295c48b46d4645627/.tmp/cf/055af941330247b59daf553a5d03faa6, store: [table=testtb-testExportFileSystemStateWithSkipTmp family=cf region=ce04d37ae25166e295c48b46d4645627] 2025-01-08T05:02:42,626 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/6a4836669ea34b387dc723203ce2f74c/.tmp/cf/090def20f49d4f809919df86a1b6a95b is 220, key is 1a1660cdcbaa081b8f7689d9250fb2c8c/cf:q/1736312561976/Put/seqid=0 2025-01-08T05:02:42,626 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/ce04d37ae25166e295c48b46d4645627/.tmp/cf/055af941330247b59daf553a5d03faa6 is 220, key is 0f7c331fb75e1648f67459b1d3b9f1da3/cf:q/1736312561974/Put/seqid=0 2025-01-08T05:02:42,638 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742383_1559 (size=15957) 2025-01-08T05:02:42,638 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742383_1559 (size=15957) 2025-01-08T05:02:42,638 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742383_1559 (size=15957) 2025-01-08T05:02:42,638 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742384_1560 (size=5526) 2025-01-08T05:02:42,638 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742384_1560 (size=5526) 2025-01-08T05:02:42,638 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742384_1560 (size=5526) 2025-01-08T05:02:43,038 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=3.2 K, hasBloomFilter=true, into tmp file hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/6a4836669ea34b387dc723203ce2f74c/.tmp/cf/090def20f49d4f809919df86a1b6a95b 2025-01-08T05:02:43,038 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=65, hasBloomFilter=true, into tmp file hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/ce04d37ae25166e295c48b46d4645627/.tmp/cf/055af941330247b59daf553a5d03faa6 2025-01-08T05:02:43,043 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/ce04d37ae25166e295c48b46d4645627/.tmp/cf/055af941330247b59daf553a5d03faa6 as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/ce04d37ae25166e295c48b46d4645627/cf/055af941330247b59daf553a5d03faa6 2025-01-08T05:02:43,043 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/6a4836669ea34b387dc723203ce2f74c/.tmp/cf/090def20f49d4f809919df86a1b6a95b as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/6a4836669ea34b387dc723203ce2f74c/cf/090def20f49d4f809919df86a1b6a95b 2025-01-08T05:02:43,047 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/6a4836669ea34b387dc723203ce2f74c/cf/090def20f49d4f809919df86a1b6a95b, entries=49, sequenceid=6, filesize=15.6 K 2025-01-08T05:02:43,047 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/ce04d37ae25166e295c48b46d4645627/cf/055af941330247b59daf553a5d03faa6, entries=1, sequenceid=6, filesize=5.4 K 2025-01-08T05:02:43,048 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.HRegion(3040): Finished flush of dataSize ~65 B/65, heapSize ~384 B/384, currentSize=0 B/0 for ce04d37ae25166e295c48b46d4645627 in 865ms, sequenceid=6, compaction requested=false 2025-01-08T05:02:43,048 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.HRegion(3040): Finished flush of dataSize ~3.19 KB/3271, heapSize ~7.13 KB/7296, currentSize=0 B/0 for 6a4836669ea34b387dc723203ce2f74c in 865ms, sequenceid=6, compaction requested=false 2025-01-08T05:02:43,048 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.HRegion(2538): Flush status journal for 6a4836669ea34b387dc723203ce2f74c: 2025-01-08T05:02:43,048 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.HRegion(2538): Flush status journal for ce04d37ae25166e295c48b46d4645627: 2025-01-08T05:02:43,048 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c. for snaptb0-testExportFileSystemStateWithSkipTmp completed. 2025-01-08T05:02:43,048 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627. for snaptb0-testExportFileSystemStateWithSkipTmp completed. 2025-01-08T05:02:43,048 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c.' region-info for snapshot=snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T05:02:43,048 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T05:02:43,048 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627.' region-info for snapshot=snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T05:02:43,048 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/6a4836669ea34b387dc723203ce2f74c/cf/090def20f49d4f809919df86a1b6a95b] hfiles 2025-01-08T05:02:43,048 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T05:02:43,048 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/6a4836669ea34b387dc723203ce2f74c/cf/090def20f49d4f809919df86a1b6a95b for snapshot=snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T05:02:43,048 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/ce04d37ae25166e295c48b46d4645627/cf/055af941330247b59daf553a5d03faa6] hfiles 2025-01-08T05:02:43,048 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/ce04d37ae25166e295c48b46d4645627/cf/055af941330247b59daf553a5d03faa6 for snapshot=snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T05:02:43,054 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742385_1561 (size=121) 2025-01-08T05:02:43,055 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742385_1561 (size=121) 2025-01-08T05:02:43,055 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742385_1561 (size=121) 2025-01-08T05:02:43,055 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742386_1562 (size=121) 2025-01-08T05:02:43,055 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742386_1562 (size=121) 2025-01-08T05:02:43,055 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742386_1562 (size=121) 2025-01-08T05:02:43,055 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c. 2025-01-08T05:02:43,055 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=208 2025-01-08T05:02:43,056 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627. 2025-01-08T05:02:43,056 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/f28094b72178:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=207 2025-01-08T05:02:43,056 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=208 2025-01-08T05:02:43,056 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportFileSystemStateWithSkipTmp on region 6a4836669ea34b387dc723203ce2f74c 2025-01-08T05:02:43,056 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.HMaster(4121): Remote procedure done, pid=207 2025-01-08T05:02:43,056 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=208, ppid=206, state=RUNNABLE; SnapshotRegionProcedure 6a4836669ea34b387dc723203ce2f74c 2025-01-08T05:02:43,056 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportFileSystemStateWithSkipTmp on region ce04d37ae25166e295c48b46d4645627 2025-01-08T05:02:43,056 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=207, ppid=206, state=RUNNABLE; SnapshotRegionProcedure ce04d37ae25166e295c48b46d4645627 2025-01-08T05:02:43,058 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=208, ppid=206, state=SUCCESS; SnapshotRegionProcedure 6a4836669ea34b387dc723203ce2f74c in 1.0270 sec 2025-01-08T05:02:43,058 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=207, resume processing ppid=206 2025-01-08T05:02:43,058 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=207, ppid=206, state=SUCCESS; SnapshotRegionProcedure ce04d37ae25166e295c48b46d4645627 in 1.0270 sec 2025-01-08T05:02:43,058 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T05:02:43,059 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T05:02:43,060 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T05:02:43,060 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T05:02:43,060 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:02:43,062 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/c4ca4238a0b923820dcc509a6f75849b202501089a38c686adc2416d83a16cba30081c05_6a4836669ea34b387dc723203ce2f74c, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/d41d8cd98f00b204e9800998ecf8427e202501088c51d0a25a30410fbe457a2e0b1c281c_ce04d37ae25166e295c48b46d4645627] hfiles 2025-01-08T05:02:43,062 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/c4ca4238a0b923820dcc509a6f75849b202501089a38c686adc2416d83a16cba30081c05_6a4836669ea34b387dc723203ce2f74c 2025-01-08T05:02:43,062 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/d41d8cd98f00b204e9800998ecf8427e202501088c51d0a25a30410fbe457a2e0b1c281c_ce04d37ae25166e295c48b46d4645627 2025-01-08T05:02:43,084 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742387_1563 (size=305) 2025-01-08T05:02:43,084 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742387_1563 (size=305) 2025-01-08T05:02:43,084 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742387_1563 (size=305) 2025-01-08T05:02:43,127 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=206 2025-01-08T05:02:43,485 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T05:02:43,485 DEBUG [PEWorker-5 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T05:02:43,486 DEBUG [PEWorker-5 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T05:02:43,492 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742388_1564 (size=1007) 2025-01-08T05:02:43,493 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742388_1564 (size=1007) 2025-01-08T05:02:43,493 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742388_1564 (size=1007) 2025-01-08T05:02:43,495 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T05:02:43,499 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T05:02:43,499 DEBUG [PEWorker-5 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemStateWithSkipTmp to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T05:02:43,500 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T05:02:43,500 DEBUG [PEWorker-5 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 }, snapshot procedure id = 206 2025-01-08T05:02:43,501 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=206, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } in 1.4820 sec 2025-01-08T05:02:44,128 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=206 2025-01-08T05:02:44,128 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportFileSystemStateWithSkipTmp, procId: 206 completed 2025-01-08T05:02:44,128 INFO [Time-limited test {}] snapshot.TestExportSnapshot(468): HDFS export destination path: hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312564128 2025-01-08T05:02:44,129 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=hdfs://localhost:45939, tgtDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312564128, rawTgtDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312564128, srcFsUri=hdfs://localhost:45939, srcDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:02:44,171 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:45939, inputRoot=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49 2025-01-08T05:02:44,171 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1149249371_22, ugi=jenkins (auth:SIMPLE)]], outputRoot=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312564128, skipTmp=true, initialOutputSnapshotDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312564128/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T05:02:44,173 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T05:02:44,178 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312564128/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T05:02:44,199 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742389_1565 (size=1007) 2025-01-08T05:02:44,199 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742389_1565 (size=1007) 2025-01-08T05:02:44,200 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742389_1565 (size=1007) 2025-01-08T05:02:44,203 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742390_1566 (size=198) 2025-01-08T05:02:44,204 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742390_1566 (size=198) 2025-01-08T05:02:44,204 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742390_1566 (size=198) 2025-01-08T05:02:44,206 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.7.0-SNAPSHOT.jar 2025-01-08T05:02:44,206 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T05:02:44,206 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T05:02:44,207 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T05:02:44,423 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0009_000001 (auth:SIMPLE) from 127.0.0.1:55994 2025-01-08T05:02:44,438 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_0/usercache/jenkins/appcache/application_1736312298603_0009/container_1736312298603_0009_01_000001/launch_container.sh] 2025-01-08T05:02:44,438 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_0/usercache/jenkins/appcache/application_1736312298603_0009/container_1736312298603_0009_01_000001/container_tokens] 2025-01-08T05:02:44,438 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-0_0/usercache/jenkins/appcache/application_1736312298603_0009/container_1736312298603_0009_01_000001/sysfs] 2025-01-08T05:02:45,449 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop-7686129554134544208.jar 2025-01-08T05:02:45,450 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T05:02:45,450 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T05:02:45,527 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop-7961703734578002655.jar 2025-01-08T05:02:45,527 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T05:02:45,528 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T05:02:45,528 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T05:02:45,528 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T05:02:45,528 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T05:02:45,528 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T05:02:45,529 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T05:02:45,529 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T05:02:45,529 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T05:02:45,529 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T05:02:45,529 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T05:02:45,529 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T05:02:45,530 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T05:02:45,530 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T05:02:45,530 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T05:02:45,530 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T05:02:45,531 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T05:02:45,531 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T05:02:45,531 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:02:45,532 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:02:45,532 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T05:02:45,532 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:02:45,532 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T05:02:45,533 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T05:02:45,533 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T05:02:45,597 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742391_1567 (size=127628) 2025-01-08T05:02:45,598 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742391_1567 (size=127628) 2025-01-08T05:02:45,598 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742391_1567 (size=127628) 2025-01-08T05:02:45,652 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742392_1568 (size=2172137) 2025-01-08T05:02:45,652 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742392_1568 (size=2172137) 2025-01-08T05:02:45,653 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742392_1568 (size=2172137) 2025-01-08T05:02:45,660 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742393_1569 (size=213228) 2025-01-08T05:02:45,660 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742393_1569 (size=213228) 2025-01-08T05:02:45,661 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742393_1569 (size=213228) 2025-01-08T05:02:45,674 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742394_1570 (size=1877034) 2025-01-08T05:02:45,674 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742394_1570 (size=1877034) 2025-01-08T05:02:45,674 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742394_1570 (size=1877034) 2025-01-08T05:02:45,683 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742395_1571 (size=533455) 2025-01-08T05:02:45,684 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742395_1571 (size=533455) 2025-01-08T05:02:45,684 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742395_1571 (size=533455) 2025-01-08T05:02:45,709 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742396_1572 (size=7280644) 2025-01-08T05:02:45,710 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742396_1572 (size=7280644) 2025-01-08T05:02:45,710 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742396_1572 (size=7280644) 2025-01-08T05:02:45,729 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742397_1573 (size=4188619) 2025-01-08T05:02:45,730 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742397_1573 (size=4188619) 2025-01-08T05:02:45,730 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742397_1573 (size=4188619) 2025-01-08T05:02:45,736 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742398_1574 (size=20406) 2025-01-08T05:02:45,736 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742398_1574 (size=20406) 2025-01-08T05:02:45,737 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742398_1574 (size=20406) 2025-01-08T05:02:45,743 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742399_1575 (size=75495) 2025-01-08T05:02:45,744 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742399_1575 (size=75495) 2025-01-08T05:02:45,744 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742399_1575 (size=75495) 2025-01-08T05:02:45,745 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T05:02:45,772 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742400_1576 (size=45609) 2025-01-08T05:02:45,772 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742400_1576 (size=45609) 2025-01-08T05:02:45,773 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742400_1576 (size=45609) 2025-01-08T05:02:45,789 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742401_1577 (size=110084) 2025-01-08T05:02:45,789 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742401_1577 (size=110084) 2025-01-08T05:02:45,789 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742401_1577 (size=110084) 2025-01-08T05:02:45,799 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742402_1578 (size=1323991) 2025-01-08T05:02:45,799 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742402_1578 (size=1323991) 2025-01-08T05:02:45,799 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742402_1578 (size=1323991) 2025-01-08T05:02:45,805 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742403_1579 (size=23076) 2025-01-08T05:02:45,805 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742403_1579 (size=23076) 2025-01-08T05:02:45,806 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742403_1579 (size=23076) 2025-01-08T05:02:45,814 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742404_1580 (size=451756) 2025-01-08T05:02:45,814 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742404_1580 (size=451756) 2025-01-08T05:02:45,814 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742404_1580 (size=451756) 2025-01-08T05:02:45,820 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742405_1581 (size=126803) 2025-01-08T05:02:45,820 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742405_1581 (size=126803) 2025-01-08T05:02:45,820 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742405_1581 (size=126803) 2025-01-08T05:02:45,827 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742406_1582 (size=322274) 2025-01-08T05:02:45,827 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742406_1582 (size=322274) 2025-01-08T05:02:45,827 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742406_1582 (size=322274) 2025-01-08T05:02:45,844 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742407_1583 (size=1832290) 2025-01-08T05:02:45,844 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742407_1583 (size=1832290) 2025-01-08T05:02:45,845 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742407_1583 (size=1832290) 2025-01-08T05:02:45,856 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742408_1584 (size=30081) 2025-01-08T05:02:45,857 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742408_1584 (size=30081) 2025-01-08T05:02:45,857 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742408_1584 (size=30081) 2025-01-08T05:02:45,868 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742409_1585 (size=53616) 2025-01-08T05:02:45,868 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742409_1585 (size=53616) 2025-01-08T05:02:45,869 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742409_1585 (size=53616) 2025-01-08T05:02:45,883 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742410_1586 (size=29229) 2025-01-08T05:02:45,883 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742410_1586 (size=29229) 2025-01-08T05:02:45,884 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742410_1586 (size=29229) 2025-01-08T05:02:45,896 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742411_1587 (size=169089) 2025-01-08T05:02:45,896 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742411_1587 (size=169089) 2025-01-08T05:02:45,897 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742411_1587 (size=169089) 2025-01-08T05:02:45,924 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742412_1588 (size=5175431) 2025-01-08T05:02:45,924 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742412_1588 (size=5175431) 2025-01-08T05:02:45,925 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742412_1588 (size=5175431) 2025-01-08T05:02:45,936 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742413_1589 (size=136454) 2025-01-08T05:02:45,936 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742413_1589 (size=136454) 2025-01-08T05:02:45,937 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742413_1589 (size=136454) 2025-01-08T05:02:45,951 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742414_1590 (size=907470) 2025-01-08T05:02:45,951 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742414_1590 (size=907470) 2025-01-08T05:02:45,951 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742414_1590 (size=907470) 2025-01-08T05:02:45,979 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742415_1591 (size=3317408) 2025-01-08T05:02:45,983 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742415_1591 (size=3317408) 2025-01-08T05:02:45,983 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742415_1591 (size=3317408) 2025-01-08T05:02:46,015 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742416_1592 (size=6350712) 2025-01-08T05:02:46,015 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742416_1592 (size=6350712) 2025-01-08T05:02:46,015 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742416_1592 (size=6350712) 2025-01-08T05:02:46,424 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742417_1593 (size=503880) 2025-01-08T05:02:46,425 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742417_1593 (size=503880) 2025-01-08T05:02:46,425 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742417_1593 (size=503880) 2025-01-08T05:02:46,443 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742418_1594 (size=4695811) 2025-01-08T05:02:46,444 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742418_1594 (size=4695811) 2025-01-08T05:02:46,444 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742418_1594 (size=4695811) 2025-01-08T05:02:46,445 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T05:02:46,447 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'snaptb0-testExportFileSystemStateWithSkipTmp' hfile list 2025-01-08T05:02:46,449 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(751): export split=0 size=33.9 K 2025-01-08T05:02:46,456 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742419_1595 (size=770) 2025-01-08T05:02:46,456 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742419_1595 (size=770) 2025-01-08T05:02:46,456 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742419_1595 (size=770) 2025-01-08T05:02:46,470 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742420_1596 (size=15) 2025-01-08T05:02:46,471 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742420_1596 (size=15) 2025-01-08T05:02:46,471 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742420_1596 (size=15) 2025-01-08T05:02:46,882 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742421_1597 (size=305049) 2025-01-08T05:02:46,882 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742421_1597 (size=305049) 2025-01-08T05:02:46,882 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742421_1597 (size=305049) 2025-01-08T05:02:46,901 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T05:02:46,901 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T05:02:47,185 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0010_000001 (auth:SIMPLE) from 127.0.0.1:43874 2025-01-08T05:02:47,422 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 5881ce671bd325587aadcdae5691632b, had cached 0 bytes from a total of 13275 2025-01-08T05:02:47,423 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region bf5538cd1c1443d97a8c76cd68b482f8, had cached 0 bytes from a total of 7282 2025-01-08T05:02:48,345 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T05:02:48,345 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemStateWithSkipTmp Metrics about Tables on a single HBase RegionServer 2025-01-08T05:02:48,345 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportWithChecksum 2025-01-08T05:02:53,066 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0010_000001 (auth:SIMPLE) from 127.0.0.1:35216 2025-01-08T05:02:53,396 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742422_1598 (size=350747) 2025-01-08T05:02:53,397 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742422_1598 (size=350747) 2025-01-08T05:02:53,397 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742422_1598 (size=350747) 2025-01-08T05:02:53,853 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T05:02:55,406 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0010_000001 (auth:SIMPLE) from 127.0.0.1:49908 2025-01-08T05:02:59,158 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742423_1599 (size=15957) 2025-01-08T05:02:59,162 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742423_1599 (size=15957) 2025-01-08T05:02:59,162 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742423_1599 (size=15957) 2025-01-08T05:02:59,311 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742424_1600 (size=8311) 2025-01-08T05:02:59,311 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742424_1600 (size=8311) 2025-01-08T05:02:59,311 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742424_1600 (size=8311) 2025-01-08T05:02:59,761 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742425_1601 (size=5526) 2025-01-08T05:02:59,761 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742425_1601 (size=5526) 2025-01-08T05:02:59,761 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742425_1601 (size=5526) 2025-01-08T05:02:59,813 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742426_1602 (size=4964) 2025-01-08T05:02:59,813 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742426_1602 (size=4964) 2025-01-08T05:02:59,813 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742426_1602 (size=4964) 2025-01-08T05:02:59,886 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742427_1603 (size=17473) 2025-01-08T05:02:59,887 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742427_1603 (size=17473) 2025-01-08T05:02:59,887 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742427_1603 (size=17473) 2025-01-08T05:02:59,900 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742428_1604 (size=476) 2025-01-08T05:02:59,901 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742428_1604 (size=476) 2025-01-08T05:02:59,902 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742428_1604 (size=476) 2025-01-08T05:02:59,938 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742429_1605 (size=17473) 2025-01-08T05:02:59,939 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742429_1605 (size=17473) 2025-01-08T05:02:59,939 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742429_1605 (size=17473) 2025-01-08T05:02:59,968 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742430_1606 (size=350747) 2025-01-08T05:02:59,971 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742430_1606 (size=350747) 2025-01-08T05:02:59,971 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742430_1606 (size=350747) 2025-01-08T05:02:59,984 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736312298603_0010_000001 (auth:SIMPLE) from 127.0.0.1:49910 2025-01-08T05:03:01,088 INFO [Time-limited test {}] snapshot.ExportSnapshot(1207): Finalize the Snapshot Export 2025-01-08T05:03:01,088 INFO [Time-limited test {}] snapshot.ExportSnapshot(1218): Verify the exported snapshot's expiration status and integrity. 2025-01-08T05:03:01,109 INFO [Time-limited test {}] snapshot.ExportSnapshot(1224): Export Completed: snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T05:03:01,110 INFO [Time-limited test {}] snapshot.TestExportSnapshot(362): Exported snapshot 2025-01-08T05:03:01,110 INFO [Time-limited test {}] snapshot.TestExportSnapshot(373): Verified filesystem state 2025-01-08T05:03:01,110 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1149249371_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp at hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T05:03:01,111 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp/.snapshotinfo 2025-01-08T05:03:01,111 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp/data.manifest 2025-01-08T05:03:01,111 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1149249371_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312564128/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp at hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312564128/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T05:03:01,111 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312564128/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp/.snapshotinfo 2025-01-08T05:03:01,111 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/export-test/export-1736312564128/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp/data.manifest 2025-01-08T05:03:01,118 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T05:03:01,118 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.3 disable testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T05:03:01,123 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=209, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T05:03:01,127 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=209 2025-01-08T05:03:01,127 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithSkipTmp","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312581127"}]},"ts":"1736312581127"} 2025-01-08T05:03:01,132 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithSkipTmp, state=DISABLING in hbase:meta 2025-01-08T05:03:01,135 INFO [PEWorker-3 {}] procedure.DisableTableProcedure(284): Set testtb-testExportFileSystemStateWithSkipTmp to state=DISABLING 2025-01-08T05:03:01,135 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=210, ppid=209, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testExportFileSystemStateWithSkipTmp}] 2025-01-08T05:03:01,137 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=211, ppid=210, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=ce04d37ae25166e295c48b46d4645627, UNASSIGN}, {pid=212, ppid=210, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=6a4836669ea34b387dc723203ce2f74c, UNASSIGN}] 2025-01-08T05:03:01,137 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=211, ppid=210, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=ce04d37ae25166e295c48b46d4645627, UNASSIGN 2025-01-08T05:03:01,137 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=212, ppid=210, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=6a4836669ea34b387dc723203ce2f74c, UNASSIGN 2025-01-08T05:03:01,138 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=211 updating hbase:meta row=ce04d37ae25166e295c48b46d4645627, regionState=CLOSING, regionLocation=f28094b72178,37967,1736312288941 2025-01-08T05:03:01,138 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=212 updating hbase:meta row=6a4836669ea34b387dc723203ce2f74c, regionState=CLOSING, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T05:03:01,140 DEBUG [PEWorker-1 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T05:03:01,140 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=213, ppid=212, state=RUNNABLE; CloseRegionProcedure 6a4836669ea34b387dc723203ce2f74c, server=f28094b72178,37583,1736312288873}] 2025-01-08T05:03:01,140 DEBUG [PEWorker-2 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T05:03:01,142 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=214, ppid=211, state=RUNNABLE; CloseRegionProcedure ce04d37ae25166e295c48b46d4645627, server=f28094b72178,37967,1736312288941}] 2025-01-08T05:03:01,228 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=209 2025-01-08T05:03:01,294 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T05:03:01,294 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,37967,1736312288941 2025-01-08T05:03:01,295 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] handler.UnassignRegionHandler(124): Close ce04d37ae25166e295c48b46d4645627 2025-01-08T05:03:01,295 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] handler.UnassignRegionHandler(124): Close 6a4836669ea34b387dc723203ce2f74c 2025-01-08T05:03:01,295 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T05:03:01,295 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T05:03:01,295 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] regionserver.HRegion(1681): Closing ce04d37ae25166e295c48b46d4645627, disabling compactions & flushes 2025-01-08T05:03:01,295 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] regionserver.HRegion(1681): Closing 6a4836669ea34b387dc723203ce2f74c, disabling compactions & flushes 2025-01-08T05:03:01,295 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627. 2025-01-08T05:03:01,295 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c. 2025-01-08T05:03:01,295 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627. 2025-01-08T05:03:01,295 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c. 2025-01-08T05:03:01,295 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627. after waiting 0 ms 2025-01-08T05:03:01,295 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c. after waiting 0 ms 2025-01-08T05:03:01,295 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627. 2025-01-08T05:03:01,295 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c. 2025-01-08T05:03:01,299 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/ce04d37ae25166e295c48b46d4645627/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T05:03:01,300 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:03:01,300 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627. 2025-01-08T05:03:01,300 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] regionserver.HRegion(1635): Region close journal for ce04d37ae25166e295c48b46d4645627: 2025-01-08T05:03:01,302 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] handler.UnassignRegionHandler(170): Closed ce04d37ae25166e295c48b46d4645627 2025-01-08T05:03:01,302 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=211 updating hbase:meta row=ce04d37ae25166e295c48b46d4645627, regionState=CLOSED 2025-01-08T05:03:01,303 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/6a4836669ea34b387dc723203ce2f74c/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T05:03:01,303 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:03:01,303 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c. 2025-01-08T05:03:01,303 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] regionserver.HRegion(1635): Region close journal for 6a4836669ea34b387dc723203ce2f74c: 2025-01-08T05:03:01,305 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] handler.UnassignRegionHandler(170): Closed 6a4836669ea34b387dc723203ce2f74c 2025-01-08T05:03:01,305 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=212 updating hbase:meta row=6a4836669ea34b387dc723203ce2f74c, regionState=CLOSED 2025-01-08T05:03:01,305 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=214, resume processing ppid=211 2025-01-08T05:03:01,306 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=214, ppid=211, state=SUCCESS; CloseRegionProcedure ce04d37ae25166e295c48b46d4645627, server=f28094b72178,37967,1736312288941 in 163 msec 2025-01-08T05:03:01,307 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=211, ppid=210, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=ce04d37ae25166e295c48b46d4645627, UNASSIGN in 168 msec 2025-01-08T05:03:01,307 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=213, resume processing ppid=212 2025-01-08T05:03:01,307 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=213, ppid=212, state=SUCCESS; CloseRegionProcedure 6a4836669ea34b387dc723203ce2f74c, server=f28094b72178,37583,1736312288873 in 166 msec 2025-01-08T05:03:01,310 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=212, resume processing ppid=210 2025-01-08T05:03:01,310 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=212, ppid=210, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=6a4836669ea34b387dc723203ce2f74c, UNASSIGN in 170 msec 2025-01-08T05:03:01,312 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=210, resume processing ppid=209 2025-01-08T05:03:01,312 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=210, ppid=209, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testExportFileSystemStateWithSkipTmp in 175 msec 2025-01-08T05:03:01,312 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithSkipTmp","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736312581312"}]},"ts":"1736312581312"} 2025-01-08T05:03:01,313 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithSkipTmp, state=DISABLED in hbase:meta 2025-01-08T05:03:01,315 INFO [PEWorker-4 {}] procedure.DisableTableProcedure(296): Set testtb-testExportFileSystemStateWithSkipTmp to state=DISABLED 2025-01-08T05:03:01,317 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=209, state=SUCCESS; DisableTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp in 198 msec 2025-01-08T05:03:01,429 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=209 2025-01-08T05:03:01,429 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testExportFileSystemStateWithSkipTmp, procId: 209 completed 2025-01-08T05:03:01,429 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.3 delete testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T05:03:01,430 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] procedure2.ProcedureExecutor(1098): Stored pid=215, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T05:03:01,431 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=215, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T05:03:01,431 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T05:03:01,432 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=215, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T05:03:01,433 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37583 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T05:03:01,435 DEBUG [HFileArchiver-22 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/ce04d37ae25166e295c48b46d4645627 2025-01-08T05:03:01,435 DEBUG [HFileArchiver-23 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/6a4836669ea34b387dc723203ce2f74c 2025-01-08T05:03:01,439 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T05:03:01,439 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T05:03:01,439 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T05:03:01,439 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T05:03:01,441 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithSkipTmp with data PBUF 2025-01-08T05:03:01,441 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithSkipTmp with data PBUF 2025-01-08T05:03:01,441 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T05:03:01,441 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T05:03:01,441 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:03:01,441 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:03:01,441 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:03:01,441 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T05:03:01,441 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T05:03:01,441 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithSkipTmp with data PBUF 2025-01-08T05:03:01,442 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=215 2025-01-08T05:03:01,442 DEBUG [HFileArchiver-23 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/6a4836669ea34b387dc723203ce2f74c/cf, FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/6a4836669ea34b387dc723203ce2f74c/recovered.edits] 2025-01-08T05:03:01,446 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithSkipTmp with data null 2025-01-08T05:03:01,446 INFO [zk-permission-watcher-pool-0 {}] access.AuthManager(135): Skipping permission cache refresh because writable data is empty 2025-01-08T05:03:01,447 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:03:01,447 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:03:01,448 DEBUG [HFileArchiver-22 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/ce04d37ae25166e295c48b46d4645627/cf, FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/ce04d37ae25166e295c48b46d4645627/recovered.edits] 2025-01-08T05:03:01,449 DEBUG [HFileArchiver-23 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/6a4836669ea34b387dc723203ce2f74c/cf/090def20f49d4f809919df86a1b6a95b to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportFileSystemStateWithSkipTmp/6a4836669ea34b387dc723203ce2f74c/cf/090def20f49d4f809919df86a1b6a95b 2025-01-08T05:03:01,450 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:03:01,451 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T05:03:01,453 DEBUG [HFileArchiver-23 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/6a4836669ea34b387dc723203ce2f74c/recovered.edits/9.seqid to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportFileSystemStateWithSkipTmp/6a4836669ea34b387dc723203ce2f74c/recovered.edits/9.seqid 2025-01-08T05:03:01,454 DEBUG [HFileArchiver-22 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/ce04d37ae25166e295c48b46d4645627/cf/055af941330247b59daf553a5d03faa6 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportFileSystemStateWithSkipTmp/ce04d37ae25166e295c48b46d4645627/cf/055af941330247b59daf553a5d03faa6 2025-01-08T05:03:01,454 DEBUG [HFileArchiver-23 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/6a4836669ea34b387dc723203ce2f74c 2025-01-08T05:03:01,457 DEBUG [HFileArchiver-22 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/ce04d37ae25166e295c48b46d4645627/recovered.edits/9.seqid to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportFileSystemStateWithSkipTmp/ce04d37ae25166e295c48b46d4645627/recovered.edits/9.seqid 2025-01-08T05:03:01,457 DEBUG [HFileArchiver-22 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testtb-testExportFileSystemStateWithSkipTmp/ce04d37ae25166e295c48b46d4645627 2025-01-08T05:03:01,457 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(313): Archived testtb-testExportFileSystemStateWithSkipTmp regions 2025-01-08T05:03:01,457 DEBUG [PEWorker-3 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a 2025-01-08T05:03:01,458 DEBUG [PEWorker-3 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf] 2025-01-08T05:03:01,462 DEBUG [PEWorker-3 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/c4ca4238a0b923820dcc509a6f75849b202501089a38c686adc2416d83a16cba30081c05_6a4836669ea34b387dc723203ce2f74c to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/c4ca4238a0b923820dcc509a6f75849b202501089a38c686adc2416d83a16cba30081c05_6a4836669ea34b387dc723203ce2f74c 2025-01-08T05:03:01,469 DEBUG [PEWorker-3 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/d41d8cd98f00b204e9800998ecf8427e202501088c51d0a25a30410fbe457a2e0b1c281c_ce04d37ae25166e295c48b46d4645627 to hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/archive/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/d41d8cd98f00b204e9800998ecf8427e202501088c51d0a25a30410fbe457a2e0b1c281c_ce04d37ae25166e295c48b46d4645627 2025-01-08T05:03:01,469 DEBUG [PEWorker-3 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a 2025-01-08T05:03:01,475 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=215, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T05:03:01,477 WARN [PEWorker-3 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testtb-testExportFileSystemStateWithSkipTmp from hbase:meta 2025-01-08T05:03:01,479 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testExportFileSystemStateWithSkipTmp' descriptor. 2025-01-08T05:03:01,481 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=215, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T05:03:01,481 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testExportFileSystemStateWithSkipTmp' from region states. 2025-01-08T05:03:01,481 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736312581481"}]},"ts":"9223372036854775807"} 2025-01-08T05:03:01,481 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736312581481"}]},"ts":"9223372036854775807"} 2025-01-08T05:03:01,486 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T05:03:01,486 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => ce04d37ae25166e295c48b46d4645627, NAME => 'testtb-testExportFileSystemStateWithSkipTmp,,1736312560512.ce04d37ae25166e295c48b46d4645627.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => 6a4836669ea34b387dc723203ce2f74c, NAME => 'testtb-testExportFileSystemStateWithSkipTmp,1,1736312560512.6a4836669ea34b387dc723203ce2f74c.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T05:03:01,486 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testExportFileSystemStateWithSkipTmp' as deleted. 2025-01-08T05:03:01,486 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithSkipTmp","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736312581486"}]},"ts":"9223372036854775807"} 2025-01-08T05:03:01,490 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testExportFileSystemStateWithSkipTmp state from META 2025-01-08T05:03:01,493 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(133): Finished pid=215, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T05:03:01,494 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=215, state=SUCCESS; DeleteTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp in 64 msec 2025-01-08T05:03:01,543 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=215 2025-01-08T05:03:01,543 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testExportFileSystemStateWithSkipTmp, procId: 215 completed 2025-01-08T05:03:01,549 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "emptySnaptb0-testExportFileSystemStateWithSkipTmp" 2025-01-08T05:03:01,552 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotManager(380): Deleting snapshot: emptySnaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T05:03:01,554 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "snaptb0-testExportFileSystemStateWithSkipTmp" 2025-01-08T05:03:01,555 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T05:03:01,580 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: snapshot.TestMobSecureExportSnapshot#testExportFileSystemStateWithSkipTmp Thread=803 (was 802) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:45665 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1826669791_1 at /127.0.0.1:50322 [Waiting for operation #3] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1149249371_22 at /127.0.0.1:53718 [Waiting for operation #3] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-23 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-22 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x168697cd-shared-pool-49 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1149249371_22 at /127.0.0.1:45280 [Waiting for operation #5] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x168697cd-shared-pool-51 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x168697cd-shared-pool-52 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Thread-8024 java.base@17.0.11/java.io.FileInputStream.readBytes(Native Method) java.base@17.0.11/java.io.FileInputStream.read(FileInputStream.java:276) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:282) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) java.base@17.0.11/sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:281) java.base@17.0.11/sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:324) java.base@17.0.11/sun.nio.cs.StreamDecoder.read(StreamDecoder.java:189) java.base@17.0.11/java.io.InputStreamReader.read(InputStreamReader.java:177) java.base@17.0.11/java.io.BufferedReader.fill(BufferedReader.java:162) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:329) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:396) app//org.apache.hadoop.util.Shell$1.run(Shell.java:1025) Potentially hanging thread: ApplicationMasterLauncher #17 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #18 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x168697cd-shared-pool-50 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: process reaper (pid 30766) java.base@17.0.11/java.lang.ProcessHandleImpl.waitForProcessExit0(Native Method) java.base@17.0.11/java.lang.ProcessHandleImpl$1.run(ProcessHandleImpl.java:150) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (366384994) connection to localhost/127.0.0.1:45665 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1149249371_22 at /127.0.0.1:50352 [Waiting for operation #4] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) - Thread LEAK? -, OpenFileDescriptor=804 (was 807), MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=712 (was 721), ProcessCount=16 (was 16), AvailableMemoryMB=1642 (was 531) - AvailableMemoryMB LEAK? - 2025-01-08T05:03:01,580 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=803 is superior to 500 2025-01-08T05:03:01,580 INFO [Time-limited test {}] hbase.HBaseTestingUtility(2861): Stopping mini mapreduce cluster... 2025-01-08T05:03:01,590 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@451087f8{node,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/node} 2025-01-08T05:03:01,593 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@306d4cf6{HTTP/1.1, (http/1.1)}{localhost:0} 2025-01-08T05:03:01,594 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2025-01-08T05:03:01,594 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@7da597ef{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/static,STOPPED} 2025-01-08T05:03:01,594 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@4578ed9f{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop.log.dir/,STOPPED} 2025-01-08T05:03:01,617 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(360): Exit code from container container_1736312298603_0010_01_000001 is : 143 2025-01-08T05:03:01,636 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-1_1/usercache/jenkins/appcache/application_1736312298603_0010/container_1736312298603_0010_01_000001/launch_container.sh] 2025-01-08T05:03:01,636 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-1_1/usercache/jenkins/appcache/application_1736312298603_0010/container_1736312298603_0010_01_000001/container_tokens] 2025-01-08T05:03:01,637 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-1_1/usercache/jenkins/appcache/application_1736312298603_0010/container_1736312298603_0010_01_000001/sysfs] 2025-01-08T05:03:05,089 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-1_1/usercache/jenkins/appcache/application_1736312298603_0010/container_1736312298603_0010_01_000002/launch_container.sh] 2025-01-08T05:03:05,089 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-1_1/usercache/jenkins/appcache/application_1736312298603_0010/container_1736312298603_0010_01_000002/container_tokens] 2025-01-08T05:03:05,089 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/MiniMRCluster_1354889031/yarn-182540655/MiniMRCluster_1354889031-localDir-nm-1_1/usercache/jenkins/appcache/application_1736312298603_0010/container_1736312298603_0010_01_000002/sysfs] 2025-01-08T05:03:05,792 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T05:03:06,828 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T05:03:08,345 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T05:03:08,805 INFO [LruBlockCacheStatsExecutor {}] hfile.LruBlockCache(1020): totalSize=880 MB, usedSize=660.99 KB, freeSize=879.35 MB, max=880 MB, blockCount=0, accesses=0, hits=0, hitRatio=0, cachingAccesses=0, cachingHits=0, cachingHitsRatio=0,evictions=29, evicted=0, evictedPerRun=0.0 2025-01-08T05:03:08,880 INFO [LruBlockCacheStatsExecutor {}] hfile.LruBlockCache(1020): totalSize=880 MB, usedSize=660.99 KB, freeSize=879.35 MB, max=880 MB, blockCount=0, accesses=0, hits=0, hitRatio=0, cachingAccesses=0, cachingHits=0, cachingHitsRatio=0,evictions=29, evicted=0, evictedPerRun=0.0 2025-01-08T05:03:08,948 INFO [LruBlockCacheStatsExecutor {}] hfile.LruBlockCache(1020): totalSize=880 MB, usedSize=662.40 KB, freeSize=879.35 MB, max=880 MB, blockCount=2, accesses=2, hits=0, hitRatio=0, cachingAccesses=2, cachingHits=0, cachingHitsRatio=0,evictions=29, evicted=0, evictedPerRun=0.0 2025-01-08T05:03:09,033 DEBUG [master/f28094b72178:0:becomeActiveMaster-MemStoreChunkPool Statistics {}] regionserver.ChunkCreator$MemStoreChunkPool$StatisticsThread(417): data stats (chunk size=2097152): current pool size=2, created chunk count=10, reused chunk count=22, reuseRatio=68.75% 2025-01-08T05:03:09,033 DEBUG [master/f28094b72178:0:becomeActiveMaster-MemStoreChunkPool Statistics {}] regionserver.ChunkCreator$MemStoreChunkPool$StatisticsThread(417): index stats (chunk size=209715): current pool size=0, created chunk count=0, reused chunk count=0, reuseRatio=0 2025-01-08T05:03:11,425 INFO [f28094b72178:39801Replication Statistics #0 {}] regionserver.Replication$ReplicationStatisticsTask(247): Global stats: WAL Edits Buffer Used=0B, Limit=268435456B 2025-01-08T05:03:11,425 INFO [f28094b72178:37967Replication Statistics #0 {}] regionserver.Replication$ReplicationStatisticsTask(247): Global stats: WAL Edits Buffer Used=0B, Limit=268435456B 2025-01-08T05:03:11,428 INFO [f28094b72178:37583Replication Statistics #0 {}] regionserver.Replication$ReplicationStatisticsTask(247): Global stats: WAL Edits Buffer Used=0B, Limit=268435456B 2025-01-08T05:03:13,324 INFO [regionserver/f28094b72178:0.Chore.1 {}] regionserver.HRegionServer$PeriodicMemStoreFlusher(2070): MemstoreFlusherChore requesting flush of hbase:meta,,1.1588230740 because 1588230740/info has an old edit so flush to free WALs after random delay 4091 ms 2025-01-08T05:03:13,847 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T05:03:14,374 INFO [regionserver/f28094b72178:0.Chore.1 {}] regionserver.HRegionServer$PeriodicMemStoreFlusher(2070): MemstoreFlusherChore requesting flush of hbase:namespace,,1736312292327.52312176a1351392d55e81604512c13b. because 52312176a1351392d55e81604512c13b/info has an old edit so flush to free WALs after random delay 191609 ms 2025-01-08T05:03:16,849 DEBUG [master/f28094b72178:0.Chore.1 {}] balancer.BaseLoadBalancer(778): Start Generate Balance plan for cluster. 2025-01-08T05:03:16,850 DEBUG [master/f28094b72178:0.Chore.1 {}] balancer.BalancerClusterState(202): Hosts are {f28094b72178=0} racks are {/default-rack=0} 2025-01-08T05:03:16,850 DEBUG [master/f28094b72178:0.Chore.1 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T05:03:16,850 DEBUG [master/f28094b72178:0.Chore.1 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T05:03:16,850 DEBUG [master/f28094b72178:0.Chore.1 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T05:03:16,850 INFO [master/f28094b72178:0.Chore.1 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T05:03:16,851 INFO [master/f28094b72178:0.Chore.1 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T05:03:16,851 INFO [master/f28094b72178:0.Chore.1 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T05:03:16,851 DEBUG [master/f28094b72178:0.Chore.1 {}] balancer.BalancerClusterState(319): Number of tables=4, number of hosts=1, number of racks=1 2025-01-08T05:03:16,852 INFO [master/f28094b72178:0.Chore.1 {}] balancer.StochasticLoadBalancer(391): Running balancer because cluster has sloppy server(s). function cost=RegionCountSkewCostFunction : (multiplier=500.0, imbalance=0.25000000000000006, need balance); PrimaryRegionCountSkewCostFunction : (not needed); MoveCostFunction : (multiplier=7.0, imbalance=0.0); ServerLocalityCostFunction : (multiplier=25.0, imbalance=0.0); RackLocalityCostFunction : (multiplier=15.0, imbalance=0.0); TableSkewCostFunction : (multiplier=35.0, imbalance=0.0); RegionReplicaHostCostFunction : (not needed); RegionReplicaRackCostFunction : (not needed); ReadRequestCostFunction : (multiplier=5.0, imbalance=1.0, need balance); WriteRequestCostFunction : (multiplier=5.0, imbalance=1.0, need balance); MemStoreSizeCostFunction : (multiplier=5.0, imbalance=0.0); StoreFileCostFunction : (multiplier=5.0, imbalance=0.0); 2025-01-08T05:03:16,853 INFO [master/f28094b72178:0.Chore.1 {}] balancer.StochasticLoadBalancer(556): Start StochasticLoadBalancer.balancer, initial weighted average imbalance=0.2242524916943522, functionCost=RegionCountSkewCostFunction : (multiplier=500.0, imbalance=0.25000000000000006, need balance); PrimaryRegionCountSkewCostFunction : (not needed); MoveCostFunction : (multiplier=7.0, imbalance=0.0); ServerLocalityCostFunction : (multiplier=25.0, imbalance=0.0); RackLocalityCostFunction : (multiplier=15.0, imbalance=0.0); TableSkewCostFunction : (multiplier=35.0, imbalance=0.0); RegionReplicaHostCostFunction : (not needed); RegionReplicaRackCostFunction : (not needed); ReadRequestCostFunction : (multiplier=5.0, imbalance=1.0, need balance); WriteRequestCostFunction : (multiplier=5.0, imbalance=1.0, need balance); MemStoreSizeCostFunction : (multiplier=5.0, imbalance=0.0); StoreFileCostFunction : (multiplier=5.0, imbalance=0.0); computedMaxSteps=12000 2025-01-08T05:03:16,984 INFO [master/f28094b72178:0.Chore.1 {}] balancer.StochasticLoadBalancer(603): Finished computing new moving plan. Computation took 132 ms to try 12000 different iterations. Found a solution that moves 1 regions; Going from a computed imbalance of 0.2242524916943522 to a new imbalance of 0.01647048202420845. funtionCost=RegionCountSkewCostFunction : (multiplier=500.0, imbalance=0.0); PrimaryRegionCountSkewCostFunction : (not needed); MoveCostFunction : (multiplier=7.0, imbalance=0.2, need balance); ServerLocalityCostFunction : (multiplier=25.0, imbalance=0.0); RackLocalityCostFunction : (multiplier=15.0, imbalance=0.0); TableSkewCostFunction : (multiplier=35.0, imbalance=0.0); RegionReplicaHostCostFunction : (not needed); RegionReplicaRackCostFunction : (not needed); ReadRequestCostFunction : (multiplier=5.0, imbalance=0.8348317674215819, need balance); WriteRequestCostFunction : (multiplier=5.0, imbalance=0.8682142682931152, need balance); MemStoreSizeCostFunction : (multiplier=5.0, imbalance=0.0); StoreFileCostFunction : (multiplier=5.0, imbalance=0.0); 2025-01-08T05:03:16,984 INFO [master/f28094b72178:0.Chore.1 {}] master.HMaster(2123): Balancer plans size is 1, the balance interval is 300000 ms, and the max number regions in transition is 5 2025-01-08T05:03:16,984 INFO [master/f28094b72178:0.Chore.1 {}] master.HMaster(2128): balance hri=1588230740, source=f28094b72178,37583,1736312288873, destination=f28094b72178,39801,1736312288779 2025-01-08T05:03:16,985 DEBUG [master/f28094b72178:0.Chore.1 {}] procedure2.ProcedureExecutor(1098): Stored pid=216, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, REOPEN/MOVE 2025-01-08T05:03:16,986 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=216, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, REOPEN/MOVE 2025-01-08T05:03:16,987 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=216 updating hbase:meta row=1588230740, regionState=CLOSING, regionLocation=f28094b72178,37583,1736312288873 2025-01-08T05:03:16,988 INFO [PEWorker-1 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as f28094b72178,37583,1736312288873, state=CLOSING 2025-01-08T05:03:16,989 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2025-01-08T05:03:16,989 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2025-01-08T05:03:16,989 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2025-01-08T05:03:16,989 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2025-01-08T05:03:16,990 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T05:03:16,990 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T05:03:16,990 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T05:03:16,990 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T05:03:16,990 DEBUG [PEWorker-1 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T05:03:16,990 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=217, ppid=216, state=RUNNABLE; CloseRegionProcedure 1588230740, server=f28094b72178,37583,1736312288873}] 2025-01-08T05:03:17,143 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to f28094b72178,37583,1736312288873 2025-01-08T05:03:17,144 INFO [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] handler.UnassignRegionHandler(124): Close 1588230740 2025-01-08T05:03:17,144 DEBUG [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T05:03:17,144 DEBUG [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2025-01-08T05:03:17,144 INFO [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2025-01-08T05:03:17,144 DEBUG [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2025-01-08T05:03:17,144 DEBUG [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2025-01-08T05:03:17,144 DEBUG [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2025-01-08T05:03:17,145 INFO [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] regionserver.HRegion(2837): Flushing 1588230740 3/3 column families, dataSize=68.66 KB heapSize=109 KB 2025-01-08T05:03:17,181 DEBUG [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/1588230740/.tmp/info/44034045d34a4e8d88d9fb32641a41c7 is 173, key is testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b./info:regioninfo/1736312477475/Put/seqid=0 2025-01-08T05:03:17,198 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742431_1607 (size=15630) 2025-01-08T05:03:17,199 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742431_1607 (size=15630) 2025-01-08T05:03:17,199 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742431_1607 (size=15630) 2025-01-08T05:03:17,416 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2464): Skipping flush on hbase:meta,,1.1588230740 because closing 2025-01-08T05:03:17,599 INFO [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=60.26 KB at sequenceid=202 (bloomFilter=true), to=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/1588230740/.tmp/info/44034045d34a4e8d88d9fb32641a41c7 2025-01-08T05:03:17,642 DEBUG [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/1588230740/.tmp/rep_barrier/dd7c4172faf641a3a01308f4a38fc2b6 is 133, key is testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452763.a739dece698030682185693f1769a5eb./rep_barrier:/1736312474819/DeleteFamily/seqid=0 2025-01-08T05:03:17,677 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742432_1608 (size=8007) 2025-01-08T05:03:17,677 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742432_1608 (size=8007) 2025-01-08T05:03:17,678 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742432_1608 (size=8007) 2025-01-08T05:03:17,678 INFO [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=2.34 KB at sequenceid=202 (bloomFilter=true), to=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/1588230740/.tmp/rep_barrier/dd7c4172faf641a3a01308f4a38fc2b6 2025-01-08T05:03:17,721 DEBUG [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/1588230740/.tmp/table/1f15f7d89dac4ff98e7b6844610f1bb1 is 127, key is testtb-testExportFileSystemStateWithMergeRegion-1,,1736312452763.a739dece698030682185693f1769a5eb./table:/1736312474819/DeleteFamily/seqid=0 2025-01-08T05:03:17,735 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742433_1609 (size=8861) 2025-01-08T05:03:17,735 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742433_1609 (size=8861) 2025-01-08T05:03:17,735 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742433_1609 (size=8861) 2025-01-08T05:03:17,736 INFO [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=6.06 KB at sequenceid=202 (bloomFilter=true), to=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/1588230740/.tmp/table/1f15f7d89dac4ff98e7b6844610f1bb1 2025-01-08T05:03:17,745 DEBUG [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/1588230740/.tmp/info/44034045d34a4e8d88d9fb32641a41c7 as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/1588230740/info/44034045d34a4e8d88d9fb32641a41c7 2025-01-08T05:03:17,749 INFO [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/1588230740/info/44034045d34a4e8d88d9fb32641a41c7, entries=84, sequenceid=202, filesize=15.3 K 2025-01-08T05:03:17,750 DEBUG [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/1588230740/.tmp/rep_barrier/dd7c4172faf641a3a01308f4a38fc2b6 as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/1588230740/rep_barrier/dd7c4172faf641a3a01308f4a38fc2b6 2025-01-08T05:03:17,756 INFO [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/1588230740/rep_barrier/dd7c4172faf641a3a01308f4a38fc2b6, entries=21, sequenceid=202, filesize=7.8 K 2025-01-08T05:03:17,757 DEBUG [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/1588230740/.tmp/table/1f15f7d89dac4ff98e7b6844610f1bb1 as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/1588230740/table/1f15f7d89dac4ff98e7b6844610f1bb1 2025-01-08T05:03:17,769 INFO [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/1588230740/table/1f15f7d89dac4ff98e7b6844610f1bb1, entries=38, sequenceid=202, filesize=8.7 K 2025-01-08T05:03:17,770 INFO [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] regionserver.HRegion(3040): Finished flush of dataSize ~68.66 KB/70312, heapSize ~108.95 KB/111568, currentSize=0 B/0 for 1588230740 in 625ms, sequenceid=202, compaction requested=false 2025-01-08T05:03:17,793 DEBUG [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/1588230740/recovered.edits/205.seqid, newMaxSeqId=205, maxSeqId=1 2025-01-08T05:03:17,794 DEBUG [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:03:17,794 DEBUG [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2025-01-08T05:03:17,794 INFO [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2025-01-08T05:03:17,794 DEBUG [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] regionserver.HRegion(1635): Region close journal for 1588230740: 2025-01-08T05:03:17,794 INFO [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] regionserver.HRegionServer(3789): Adding 1588230740 move to f28094b72178,39801,1736312288779 record at close sequenceid=202 2025-01-08T05:03:17,795 INFO [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META, pid=217}] handler.UnassignRegionHandler(170): Closed 1588230740 2025-01-08T05:03:17,796 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=216 updating hbase:meta row=1588230740, regionState=CLOSED 2025-01-08T05:03:17,796 WARN [PEWorker-2 {}] zookeeper.MetaTableLocator(168): Tried to set null ServerName in hbase:meta; skipping -- ServerName required 2025-01-08T05:03:17,798 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=217, resume processing ppid=216 2025-01-08T05:03:17,798 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=217, ppid=216, state=SUCCESS; CloseRegionProcedure 1588230740, server=f28094b72178,37583,1736312288873 in 806 msec 2025-01-08T05:03:17,799 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=216, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, REOPEN/MOVE; state=CLOSED, location=f28094b72178,39801,1736312288779; forceNewPlan=false, retain=false 2025-01-08T05:03:17,949 INFO [f28094b72178:42195 {}] balancer.BaseLoadBalancer(546): Reassigned 1 regions. 1 retained the pre-restart assignment. 2025-01-08T05:03:17,949 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=216 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=f28094b72178,39801,1736312288779 2025-01-08T05:03:17,950 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as f28094b72178,39801,1736312288779, state=OPENING 2025-01-08T05:03:17,952 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2025-01-08T05:03:17,952 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2025-01-08T05:03:17,952 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2025-01-08T05:03:17,952 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2025-01-08T05:03:17,952 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=218, ppid=216, state=RUNNABLE; OpenRegionProcedure 1588230740, server=f28094b72178,39801,1736312288779}] 2025-01-08T05:03:17,952 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T05:03:17,952 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T05:03:17,952 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T05:03:17,952 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T05:03:18,104 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to f28094b72178,39801,1736312288779 2025-01-08T05:03:18,107 INFO [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=218}] handler.AssignRegionHandler(135): Open hbase:meta,,1.1588230740 2025-01-08T05:03:18,107 INFO [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=218}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2025-01-08T05:03:18,107 INFO [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=218}] monitor.StreamSlowMonitor(122): New stream slow monitor .meta 2025-01-08T05:03:18,108 INFO [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=218}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=f28094b72178%2C39801%2C1736312288779.meta, suffix=.meta, logDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/WALs/f28094b72178,39801,1736312288779, archiveDir=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/oldWALs, maxLogs=32 2025-01-08T05:03:18,124 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=218}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(617): When create output stream for /user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/WALs/f28094b72178,39801,1736312288779/f28094b72178%2C39801%2C1736312288779.meta.1736312598109.meta, exclude list is [], retry=0 2025-01-08T05:03:18,127 DEBUG [RS-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45867,DS-84b09c6f-eff8-44d2-9077-117e68175eb1,DISK] 2025-01-08T05:03:18,128 DEBUG [RS-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:40353,DS-0a2953dd-a4c6-4650-9631-02981d9d1e8d,DISK] 2025-01-08T05:03:18,128 DEBUG [RS-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:33991,DS-f676faff-d67d-404f-bb0f-b90001fc3639,DISK] 2025-01-08T05:03:18,155 INFO [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=218}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/WALs/f28094b72178,39801,1736312288779/f28094b72178%2C39801%2C1736312288779.meta.1736312598109.meta 2025-01-08T05:03:18,158 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=218}] wal.AbstractFSWAL(925): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:34673:34673),(127.0.0.1/127.0.0.1:33789:33789),(127.0.0.1/127.0.0.1:45387:45387)] 2025-01-08T05:03:18,159 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=218}] regionserver.HRegion(7285): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2025-01-08T05:03:18,159 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=218}] regionserver.HRegion(7999): Registered coprocessor service: region=hbase:meta,,1 service=AccessControlService 2025-01-08T05:03:18,160 INFO [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=218}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T05:03:18,160 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=218}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2025-01-08T05:03:18,160 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=218}] regionserver.HRegion(7999): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2025-01-08T05:03:18,160 INFO [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=218}] regionserver.RegionCoprocessorHost(436): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2025-01-08T05:03:18,160 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=218}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2025-01-08T05:03:18,160 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=218}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T05:03:18,160 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=218}] regionserver.HRegion(7327): checking encryption for 1588230740 2025-01-08T05:03:18,160 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=218}] regionserver.HRegion(7330): checking classloading for 1588230740 2025-01-08T05:03:18,171 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2025-01-08T05:03:18,178 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2025-01-08T05:03:18,178 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:03:18,195 DEBUG [StoreOpener-1588230740-1 {}] regionserver.StoreEngine(277): loaded hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/1588230740/info/44034045d34a4e8d88d9fb32641a41c7 2025-01-08T05:03:18,195 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2025-01-08T05:03:18,196 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2025-01-08T05:03:18,197 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2025-01-08T05:03:18,197 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:03:18,203 DEBUG [StoreOpener-1588230740-1 {}] regionserver.StoreEngine(277): loaded hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/1588230740/rep_barrier/dd7c4172faf641a3a01308f4a38fc2b6 2025-01-08T05:03:18,203 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2025-01-08T05:03:18,203 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2025-01-08T05:03:18,204 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2025-01-08T05:03:18,204 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T05:03:18,223 DEBUG [StoreOpener-1588230740-1 {}] regionserver.StoreEngine(277): loaded hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/1588230740/table/1f15f7d89dac4ff98e7b6844610f1bb1 2025-01-08T05:03:18,223 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2025-01-08T05:03:18,224 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=218}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/1588230740 2025-01-08T05:03:18,225 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=218}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/1588230740 2025-01-08T05:03:18,226 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=218}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (42.7 M)) instead. 2025-01-08T05:03:18,227 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=218}] regionserver.HRegion(1085): writing seq id for 1588230740 2025-01-08T05:03:18,228 INFO [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=218}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=206; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=65109900, jitterRate=-0.029786884784698486}}}, FlushLargeStoresPolicy{flushSizeLowerBound=44739242} 2025-01-08T05:03:18,228 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=218}] regionserver.HRegion(1001): Region open journal for 1588230740: 2025-01-08T05:03:18,229 INFO [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=218}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:meta,,1.1588230740, pid=218, masterSystemTime=1736312598104 2025-01-08T05:03:18,231 DEBUG [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=218}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:meta,,1.1588230740 2025-01-08T05:03:18,231 INFO [RS_OPEN_META-regionserver/f28094b72178:0-0 {event_type=M_RS_OPEN_META, pid=218}] handler.AssignRegionHandler(164): Opened hbase:meta,,1.1588230740 2025-01-08T05:03:18,231 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=216 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=206, regionLocation=f28094b72178,39801,1736312288779 2025-01-08T05:03:18,232 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as f28094b72178,39801,1736312288779, state=OPEN 2025-01-08T05:03:18,233 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2025-01-08T05:03:18,233 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2025-01-08T05:03:18,233 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2025-01-08T05:03:18,233 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2025-01-08T05:03:18,233 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T05:03:18,233 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T05:03:18,233 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T05:03:18,233 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T05:03:18,234 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=218, resume processing ppid=216 2025-01-08T05:03:18,234 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=218, ppid=216, state=SUCCESS; OpenRegionProcedure 1588230740, server=f28094b72178,39801,1736312288779 in 281 msec 2025-01-08T05:03:18,235 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=216, state=SUCCESS; TransitRegionStateProcedure table=hbase:meta, region=1588230740, REOPEN/MOVE in 1.2490 sec 2025-01-08T05:03:18,295 DEBUG [master/f28094b72178:0.Chore.1 {}] master.HMaster(2159): Balancer is going into sleep until next period in 300000ms 2025-01-08T05:03:18,303 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=37583 {}] ipc.CallRunner(138): callId: 386 service: ClientService methodName: Scan size: 123 connection: 172.17.0.3:55258 deadline: 1736312658298, exception=org.apache.hadoop.hbase.exceptions.RegionMovedException: Region moved to: hostname=f28094b72178 port=39801 startCode=1736312288779. As of locationSeqNum=202. 2025-01-08T05:03:18,409 DEBUG [hconnection-0x168697cd-shared-pool-53 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T05:03:18,411 INFO [RS-EventLoopGroup-3-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:37218, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T05:03:18,442 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(219): Skipping table testExportExpiredSnapshot because normalization is disabled in its table properties and normalization is also disabled at table level by default 2025-01-08T05:03:18,461 DEBUG [master/f28094b72178:0.Chore.1 {}] janitor.CatalogJanitor(179): 2025-01-08T05:03:18,657 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@56fd86b7{node,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/node} 2025-01-08T05:03:18,658 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@22c3ce0e{HTTP/1.1, (http/1.1)}{localhost:0} 2025-01-08T05:03:18,658 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2025-01-08T05:03:18,658 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@23dc62be{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/static,STOPPED} 2025-01-08T05:03:18,658 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@34de0adc{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop.log.dir/,STOPPED} 2025-01-08T05:03:23,296 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T05:03:24,160 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2025-01-08T05:03:24,323 INFO [regionserver/f28094b72178:0.Chore.1 {}] regionserver.HRegionServer$PeriodicMemStoreFlusher(2070): MemstoreFlusherChore requesting flush of hbase:acl,,1736312293651.c1df98ea008addbfc51d7e8682ea8fd9. because c1df98ea008addbfc51d7e8682ea8fd9/l has an old edit so flush to free WALs after random delay 219364 ms 2025-01-08T05:03:32,423 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 5881ce671bd325587aadcdae5691632b, had cached 0 bytes from a total of 13275 2025-01-08T05:03:32,423 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region bf5538cd1c1443d97a8c76cd68b482f8, had cached 0 bytes from a total of 7282 2025-01-08T05:03:35,715 ERROR [Thread[Thread-416,5,FailOnTimeoutGroup] {}] delegation.AbstractDelegationTokenSecretManager$ExpiredTokenRemover(852): ExpiredTokenRemover received java.lang.InterruptedException: sleep interrupted 2025-01-08T05:03:35,715 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@1a80546{cluster,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/cluster} 2025-01-08T05:03:35,716 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@31f02c44{HTTP/1.1, (http/1.1)}{localhost:0} 2025-01-08T05:03:35,716 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2025-01-08T05:03:35,717 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3022421d{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/static,STOPPED} 2025-01-08T05:03:35,717 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@379c9ae7{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop.log.dir/,STOPPED} 2025-01-08T05:03:35,727 WARN [ApplicationMaster Launcher {}] amlauncher.ApplicationMasterLauncher$LauncherThread(122): org.apache.hadoop.yarn.server.resourcemanager.amlauncher.ApplicationMasterLauncher$LauncherThread interrupted. Returning. 2025-01-08T05:03:35,743 ERROR [SchedulerEventDispatcher:Event Processor {}] event.EventDispatcher$EventProcessor(72): Returning, interrupted : java.lang.InterruptedException 2025-01-08T05:03:35,743 ERROR [ResourceManager Event Processor Monitor {}] resourcemanager.ResourceManager$SchedulerEventDispatcher$EventProcessorMonitor(1193): Returning, interrupted : java.lang.InterruptedException: sleep interrupted 2025-01-08T05:03:35,747 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741830_1006 (size=964715) 2025-01-08T05:03:35,751 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741830_1006 (size=964715) 2025-01-08T05:03:35,751 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741830_1006 (size=964715) 2025-01-08T05:03:35,764 ERROR [Thread[Thread-440,5,FailOnTimeoutGroup] {}] delegation.AbstractDelegationTokenSecretManager$ExpiredTokenRemover(852): ExpiredTokenRemover received java.lang.InterruptedException: sleep interrupted 2025-01-08T05:03:35,769 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@6d9e8f32{jobhistory,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/jobhistory} 2025-01-08T05:03:35,774 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@1b10f19e{HTTP/1.1, (http/1.1)}{localhost:0} 2025-01-08T05:03:35,774 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2025-01-08T05:03:35,774 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@5c5cd02{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/static,STOPPED} 2025-01-08T05:03:35,775 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@2c84573a{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop.log.dir/,STOPPED} 2025-01-08T05:03:35,781 ERROR [Thread[Thread-398,5,FailOnTimeoutGroup] {}] delegation.AbstractDelegationTokenSecretManager$ExpiredTokenRemover(852): ExpiredTokenRemover received java.lang.InterruptedException: sleep interrupted 2025-01-08T05:03:35,782 INFO [Time-limited test {}] hbase.HBaseTestingUtility(2864): Mini mapreduce cluster stopped 2025-01-08T05:03:35,782 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1340): Shutting down minicluster 2025-01-08T05:03:35,782 INFO [Time-limited test {}] client.ConnectionImplementation(2127): Closing master protocol: MasterService 2025-01-08T05:03:35,782 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x4476f33e to 127.0.0.1:59048 2025-01-08T05:03:35,782 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:03:35,783 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2025-01-08T05:03:35,783 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=470953544, stopped=false 2025-01-08T05:03:35,783 DEBUG [Time-limited test {}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:03:35,783 DEBUG [Time-limited test {}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.SecureTestUtil$MasterSyncObserver 2025-01-08T05:03:35,783 INFO [Time-limited test {}] master.ServerManager(987): Cluster shutdown requested of master=f28094b72178,42195,1736312287279 2025-01-08T05:03:35,785 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2025-01-08T05:03:35,785 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2025-01-08T05:03:35,785 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T05:03:35,785 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T05:03:35,785 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2025-01-08T05:03:35,785 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T05:03:35,785 INFO [Time-limited test {}] procedure2.ProcedureExecutor(700): Stopping 2025-01-08T05:03:35,785 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:03:35,786 INFO [Time-limited test {}] regionserver.HRegionServer(2561): ***** STOPPING region server 'f28094b72178,39801,1736312288779' ***** 2025-01-08T05:03:35,786 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2025-01-08T05:03:35,786 DEBUG [Time-limited test {}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:03:35,786 INFO [Time-limited test {}] regionserver.HRegionServer(2575): STOPPED: Shutdown requested 2025-01-08T05:03:35,786 INFO [Time-limited test {}] regionserver.HRegionServer(2561): ***** STOPPING region server 'f28094b72178,37583,1736312288873' ***** 2025-01-08T05:03:35,786 DEBUG [Time-limited test {}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:03:35,786 INFO [Time-limited test {}] regionserver.HRegionServer(2575): STOPPED: Shutdown requested 2025-01-08T05:03:35,786 INFO [Time-limited test {}] regionserver.HRegionServer(2561): ***** STOPPING region server 'f28094b72178,37967,1736312288941' ***** 2025-01-08T05:03:35,786 DEBUG [Time-limited test {}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:03:35,786 INFO [Time-limited test {}] regionserver.HRegionServer(2575): STOPPED: Shutdown requested 2025-01-08T05:03:35,786 INFO [RS:0;f28094b72178:39801 {}] regionserver.HeapMemoryManager(220): Stopping 2025-01-08T05:03:35,786 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2025-01-08T05:03:35,786 INFO [RS:2;f28094b72178:37967 {}] regionserver.HeapMemoryManager(220): Stopping 2025-01-08T05:03:35,786 INFO [RS:0;f28094b72178:39801 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2025-01-08T05:03:35,786 INFO [RS:2;f28094b72178:37967 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2025-01-08T05:03:35,786 INFO [RS:2;f28094b72178:37967 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2025-01-08T05:03:35,786 INFO [RS:0;f28094b72178:39801 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2025-01-08T05:03:35,787 INFO [RS:1;f28094b72178:37583 {}] regionserver.HeapMemoryManager(220): Stopping 2025-01-08T05:03:35,787 INFO [RS:0;f28094b72178:39801 {}] regionserver.HRegionServer(3579): Received CLOSE for 5881ce671bd325587aadcdae5691632b 2025-01-08T05:03:35,787 INFO [RS:1;f28094b72178:37583 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2025-01-08T05:03:35,787 INFO [RS:1;f28094b72178:37583 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2025-01-08T05:03:35,787 INFO [RS:1;f28094b72178:37583 {}] regionserver.HRegionServer(3579): Received CLOSE for 52312176a1351392d55e81604512c13b 2025-01-08T05:03:35,787 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(373): MemStoreFlusher.0 exiting 2025-01-08T05:03:35,787 INFO [RS:0;f28094b72178:39801 {}] regionserver.HRegionServer(1224): stopping server f28094b72178,39801,1736312288779 2025-01-08T05:03:35,787 INFO [RS:1;f28094b72178:37583 {}] regionserver.HRegionServer(3579): Received CLOSE for c1df98ea008addbfc51d7e8682ea8fd9 2025-01-08T05:03:35,787 DEBUG [RS:0;f28094b72178:39801 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:03:35,787 INFO [RS:1;f28094b72178:37583 {}] regionserver.HRegionServer(1224): stopping server f28094b72178,37583,1736312288873 2025-01-08T05:03:35,787 DEBUG [RS:1;f28094b72178:37583 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:03:35,787 INFO [RS:0;f28094b72178:39801 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2025-01-08T05:03:35,787 INFO [RS:1;f28094b72178:37583 {}] regionserver.HRegionServer(1599): Waiting on 2 regions to close 2025-01-08T05:03:35,787 INFO [RS:0;f28094b72178:39801 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2025-01-08T05:03:35,787 DEBUG [RS:1;f28094b72178:37583 {}] regionserver.HRegionServer(1603): Online Regions={52312176a1351392d55e81604512c13b=hbase:namespace,,1736312292327.52312176a1351392d55e81604512c13b., c1df98ea008addbfc51d7e8682ea8fd9=hbase:acl,,1736312293651.c1df98ea008addbfc51d7e8682ea8fd9.} 2025-01-08T05:03:35,787 INFO [RS:0;f28094b72178:39801 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2025-01-08T05:03:35,787 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(373): MemStoreFlusher.0 exiting 2025-01-08T05:03:35,787 INFO [RS:0;f28094b72178:39801 {}] regionserver.HRegionServer(3579): Received CLOSE for 1588230740 2025-01-08T05:03:35,788 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2025-01-08T05:03:35,788 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T05:03:35,788 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2025-01-08T05:03:35,788 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2025-01-08T05:03:35,788 INFO [RS:2;f28094b72178:37967 {}] regionserver.HRegionServer(3579): Received CLOSE for bf5538cd1c1443d97a8c76cd68b482f8 2025-01-08T05:03:35,789 INFO [RS:2;f28094b72178:37967 {}] regionserver.HRegionServer(1224): stopping server f28094b72178,37967,1736312288941 2025-01-08T05:03:35,789 DEBUG [RS:2;f28094b72178:37967 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:03:35,789 INFO [RS:2;f28094b72178:37967 {}] regionserver.HRegionServer(1599): Waiting on 1 regions to close 2025-01-08T05:03:35,789 DEBUG [RS:2;f28094b72178:37967 {}] regionserver.HRegionServer(1603): Online Regions={bf5538cd1c1443d97a8c76cd68b482f8=testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8.} 2025-01-08T05:03:35,790 INFO [RS:0;f28094b72178:39801 {}] regionserver.HRegionServer(1599): Waiting on 2 regions to close 2025-01-08T05:03:35,790 DEBUG [RS:0;f28094b72178:39801 {}] regionserver.HRegionServer(1603): Online Regions={5881ce671bd325587aadcdae5691632b=testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b., 1588230740=hbase:meta,,1.1588230740} 2025-01-08T05:03:35,790 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(373): MemStoreFlusher.0 exiting 2025-01-08T05:03:35,792 INFO [regionserver/f28094b72178:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2025-01-08T05:03:35,793 DEBUG [RS:1;f28094b72178:37583 {}] regionserver.HRegionServer(1629): Waiting on 52312176a1351392d55e81604512c13b, c1df98ea008addbfc51d7e8682ea8fd9 2025-01-08T05:03:35,793 DEBUG [RS:0;f28094b72178:39801 {}] regionserver.HRegionServer(1629): Waiting on 1588230740, 5881ce671bd325587aadcdae5691632b 2025-01-08T05:03:35,793 DEBUG [RS:2;f28094b72178:37967 {}] regionserver.HRegionServer(1629): Waiting on bf5538cd1c1443d97a8c76cd68b482f8 2025-01-08T05:03:35,793 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 52312176a1351392d55e81604512c13b, disabling compactions & flushes 2025-01-08T05:03:35,793 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing bf5538cd1c1443d97a8c76cd68b482f8, disabling compactions & flushes 2025-01-08T05:03:35,793 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region hbase:namespace,,1736312292327.52312176a1351392d55e81604512c13b. 2025-01-08T05:03:35,793 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8. 2025-01-08T05:03:35,793 DEBUG [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2025-01-08T05:03:35,793 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1736312292327.52312176a1351392d55e81604512c13b. 2025-01-08T05:03:35,793 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8. 2025-01-08T05:03:35,793 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1736312292327.52312176a1351392d55e81604512c13b. after waiting 0 ms 2025-01-08T05:03:35,793 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8. after waiting 0 ms 2025-01-08T05:03:35,793 INFO [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2025-01-08T05:03:35,793 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8. 2025-01-08T05:03:35,793 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1736312292327.52312176a1351392d55e81604512c13b. 2025-01-08T05:03:35,793 DEBUG [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2025-01-08T05:03:35,793 DEBUG [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2025-01-08T05:03:35,793 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(2837): Flushing 52312176a1351392d55e81604512c13b 1/1 column families, dataSize=78 B heapSize=488 B 2025-01-08T05:03:35,793 DEBUG [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2025-01-08T05:03:35,794 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T05:03:35,798 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 5881ce671bd325587aadcdae5691632b, disabling compactions & flushes 2025-01-08T05:03:35,798 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b. 2025-01-08T05:03:35,798 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b. 2025-01-08T05:03:35,798 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b. after waiting 0 ms 2025-01-08T05:03:35,798 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b. 2025-01-08T05:03:35,821 INFO [regionserver/f28094b72178:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2025-01-08T05:03:35,843 DEBUG [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/meta/1588230740/recovered.edits/208.seqid, newMaxSeqId=208, maxSeqId=205 2025-01-08T05:03:35,844 DEBUG [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:03:35,844 DEBUG [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2025-01-08T05:03:35,844 INFO [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2025-01-08T05:03:35,844 DEBUG [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1635): Region close journal for 1588230740: 2025-01-08T05:03:35,844 DEBUG [RS_CLOSE_META-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2025-01-08T05:03:35,854 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/namespace/52312176a1351392d55e81604512c13b/.tmp/info/fb2dcb1681cc4e499b3e7d7c44915ac5 is 45, key is default/info:d/1736312293494/Put/seqid=0 2025-01-08T05:03:35,856 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportExpiredSnapshot/bf5538cd1c1443d97a8c76cd68b482f8/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2025-01-08T05:03:35,857 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/default/testExportExpiredSnapshot/5881ce671bd325587aadcdae5691632b/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2025-01-08T05:03:35,857 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:03:35,857 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8. 2025-01-08T05:03:35,857 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for bf5538cd1c1443d97a8c76cd68b482f8: 2025-01-08T05:03:35,857 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed testExportExpiredSnapshot,,1736312477038.bf5538cd1c1443d97a8c76cd68b482f8. 2025-01-08T05:03:35,858 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:03:35,858 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b. 2025-01-08T05:03:35,858 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 5881ce671bd325587aadcdae5691632b: 2025-01-08T05:03:35,858 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed testExportExpiredSnapshot,1,1736312477038.5881ce671bd325587aadcdae5691632b. 2025-01-08T05:03:35,861 INFO [regionserver/f28094b72178:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2025-01-08T05:03:35,864 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742435_1611 (size=5037) 2025-01-08T05:03:35,865 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742435_1611 (size=5037) 2025-01-08T05:03:35,865 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742435_1611 (size=5037) 2025-01-08T05:03:35,866 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=78 B at sequenceid=6 (bloomFilter=true), to=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/namespace/52312176a1351392d55e81604512c13b/.tmp/info/fb2dcb1681cc4e499b3e7d7c44915ac5 2025-01-08T05:03:35,872 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/namespace/52312176a1351392d55e81604512c13b/.tmp/info/fb2dcb1681cc4e499b3e7d7c44915ac5 as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/namespace/52312176a1351392d55e81604512c13b/info/fb2dcb1681cc4e499b3e7d7c44915ac5 2025-01-08T05:03:35,877 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/namespace/52312176a1351392d55e81604512c13b/info/fb2dcb1681cc4e499b3e7d7c44915ac5, entries=2, sequenceid=6, filesize=4.9 K 2025-01-08T05:03:35,878 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(3040): Finished flush of dataSize ~78 B/78, heapSize ~472 B/472, currentSize=0 B/0 for 52312176a1351392d55e81604512c13b in 85ms, sequenceid=6, compaction requested=false 2025-01-08T05:03:35,881 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/namespace/52312176a1351392d55e81604512c13b/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T05:03:35,882 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:03:35,882 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed hbase:namespace,,1736312292327.52312176a1351392d55e81604512c13b. 2025-01-08T05:03:35,882 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 52312176a1351392d55e81604512c13b: 2025-01-08T05:03:35,882 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed hbase:namespace,,1736312292327.52312176a1351392d55e81604512c13b. 2025-01-08T05:03:35,882 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing c1df98ea008addbfc51d7e8682ea8fd9, disabling compactions & flushes 2025-01-08T05:03:35,882 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region hbase:acl,,1736312293651.c1df98ea008addbfc51d7e8682ea8fd9. 2025-01-08T05:03:35,882 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:acl,,1736312293651.c1df98ea008addbfc51d7e8682ea8fd9. 2025-01-08T05:03:35,882 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on hbase:acl,,1736312293651.c1df98ea008addbfc51d7e8682ea8fd9. after waiting 0 ms 2025-01-08T05:03:35,882 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region hbase:acl,,1736312293651.c1df98ea008addbfc51d7e8682ea8fd9. 2025-01-08T05:03:35,882 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(2837): Flushing c1df98ea008addbfc51d7e8682ea8fd9 1/1 column families, dataSize=1.38 KB heapSize=3.33 KB 2025-01-08T05:03:35,906 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/acl/c1df98ea008addbfc51d7e8682ea8fd9/.tmp/l/3995bfbd2d774fe8a6c976abc0581d23 is 74, key is testtb-testExportFileSystemStateWithMergeRegion-1/l:/1736312474798/DeleteFamily/seqid=0 2025-01-08T05:03:35,912 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742436_1612 (size=5695) 2025-01-08T05:03:35,913 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742436_1612 (size=5695) 2025-01-08T05:03:35,915 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742436_1612 (size=5695) 2025-01-08T05:03:35,921 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.38 KB at sequenceid=27 (bloomFilter=false), to=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/acl/c1df98ea008addbfc51d7e8682ea8fd9/.tmp/l/3995bfbd2d774fe8a6c976abc0581d23 2025-01-08T05:03:35,936 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.StoreFileReader(539): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for 3995bfbd2d774fe8a6c976abc0581d23 2025-01-08T05:03:35,937 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/acl/c1df98ea008addbfc51d7e8682ea8fd9/.tmp/l/3995bfbd2d774fe8a6c976abc0581d23 as hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/acl/c1df98ea008addbfc51d7e8682ea8fd9/l/3995bfbd2d774fe8a6c976abc0581d23 2025-01-08T05:03:35,942 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.StoreFileReader(539): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for 3995bfbd2d774fe8a6c976abc0581d23 2025-01-08T05:03:35,942 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/acl/c1df98ea008addbfc51d7e8682ea8fd9/l/3995bfbd2d774fe8a6c976abc0581d23, entries=12, sequenceid=27, filesize=5.6 K 2025-01-08T05:03:35,946 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(3040): Finished flush of dataSize ~1.38 KB/1412, heapSize ~3.31 KB/3392, currentSize=0 B/0 for c1df98ea008addbfc51d7e8682ea8fd9 in 63ms, sequenceid=27, compaction requested=false 2025-01-08T05:03:35,965 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/data/hbase/acl/c1df98ea008addbfc51d7e8682ea8fd9/recovered.edits/30.seqid, newMaxSeqId=30, maxSeqId=1 2025-01-08T05:03:35,965 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:03:35,965 INFO [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed hbase:acl,,1736312293651.c1df98ea008addbfc51d7e8682ea8fd9. 2025-01-08T05:03:35,965 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for c1df98ea008addbfc51d7e8682ea8fd9: 2025-01-08T05:03:35,965 DEBUG [RS_CLOSE_REGION-regionserver/f28094b72178:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed hbase:acl,,1736312293651.c1df98ea008addbfc51d7e8682ea8fd9. 2025-01-08T05:03:35,993 INFO [RS:1;f28094b72178:37583 {}] regionserver.HRegionServer(1250): stopping server f28094b72178,37583,1736312288873; all regions closed. 2025-01-08T05:03:35,993 INFO [RS:0;f28094b72178:39801 {}] regionserver.HRegionServer(1250): stopping server f28094b72178,39801,1736312288779; all regions closed. 2025-01-08T05:03:35,993 INFO [RS:2;f28094b72178:37967 {}] regionserver.HRegionServer(1250): stopping server f28094b72178,37967,1736312288941; all regions closed. 2025-01-08T05:03:36,000 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741836_1012 (size=80694) 2025-01-08T05:03:36,000 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741836_1012 (size=80694) 2025-01-08T05:03:36,002 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741836_1012 (size=80694) 2025-01-08T05:03:36,007 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741834_1010 (size=15693) 2025-01-08T05:03:36,008 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741834_1010 (size=15693) 2025-01-08T05:03:36,011 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073742434_1610 (size=662) 2025-01-08T05:03:36,011 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073742434_1610 (size=662) 2025-01-08T05:03:36,011 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741834_1010 (size=15693) 2025-01-08T05:03:36,012 DEBUG [RS:1;f28094b72178:37583 {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/oldWALs 2025-01-08T05:03:36,012 INFO [RS:1;f28094b72178:37583 {}] wal.AbstractFSWAL(1074): Closed WAL: AsyncFSWAL f28094b72178%2C37583%2C1736312288873.meta:.meta(num 1736312291976) 2025-01-08T05:03:36,013 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073742434_1610 (size=662) 2025-01-08T05:03:36,014 WARN [Close-WAL-Writer-0 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(743): complete file /user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/WALs/f28094b72178,39801,1736312288779/f28094b72178%2C39801%2C1736312288779.meta.1736312598109.meta not finished, retry = 0 2025-01-08T05:03:36,017 DEBUG [RS:2;f28094b72178:37967 {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/oldWALs 2025-01-08T05:03:36,017 INFO [RS:2;f28094b72178:37967 {}] wal.AbstractFSWAL(1074): Closed WAL: AsyncFSWAL f28094b72178%2C37967%2C1736312288941:(num 1736312291574) 2025-01-08T05:03:36,017 DEBUG [RS:2;f28094b72178:37967 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:03:36,017 INFO [RS:2;f28094b72178:37967 {}] regionserver.LeaseManager(133): Closed leases 2025-01-08T05:03:36,017 INFO [RS:2;f28094b72178:37967 {}] hbase.ChoreService(370): Chore service for: regionserver/f28094b72178:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS] on shutdown 2025-01-08T05:03:36,018 INFO [RS:2;f28094b72178:37967 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2025-01-08T05:03:36,018 INFO [RS:2;f28094b72178:37967 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2025-01-08T05:03:36,018 INFO [RS:2;f28094b72178:37967 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2025-01-08T05:03:36,018 INFO [regionserver/f28094b72178:0.logRoller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2025-01-08T05:03:36,018 INFO [RS:2;f28094b72178:37967 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.3:37967 2025-01-08T05:03:36,019 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741835_1011 (size=16083) 2025-01-08T05:03:36,019 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741835_1011 (size=16083) 2025-01-08T05:03:36,019 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741835_1011 (size=16083) 2025-01-08T05:03:36,026 DEBUG [RS:1;f28094b72178:37583 {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/oldWALs 2025-01-08T05:03:36,026 INFO [RS:1;f28094b72178:37583 {}] wal.AbstractFSWAL(1074): Closed WAL: AsyncFSWAL f28094b72178%2C37583%2C1736312288873:(num 1736312291574) 2025-01-08T05:03:36,026 DEBUG [RS:1;f28094b72178:37583 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:03:36,026 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/f28094b72178,37967,1736312288941 2025-01-08T05:03:36,026 INFO [RS:1;f28094b72178:37583 {}] regionserver.LeaseManager(133): Closed leases 2025-01-08T05:03:36,027 INFO [RS:1;f28094b72178:37583 {}] hbase.ChoreService(370): Chore service for: regionserver/f28094b72178:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS] on shutdown 2025-01-08T05:03:36,027 INFO [RS:1;f28094b72178:37583 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2025-01-08T05:03:36,027 INFO [RS:1;f28094b72178:37583 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2025-01-08T05:03:36,027 INFO [RS:1;f28094b72178:37583 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2025-01-08T05:03:36,027 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2025-01-08T05:03:36,027 INFO [RS:1;f28094b72178:37583 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.3:37583 2025-01-08T05:03:36,027 INFO [regionserver/f28094b72178:0.logRoller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2025-01-08T05:03:36,036 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/f28094b72178,37583,1736312288873 2025-01-08T05:03:36,037 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [f28094b72178,37583,1736312288873] 2025-01-08T05:03:36,037 DEBUG [RegionServerTracker-0 {}] master.DeadServer(103): Processing f28094b72178,37583,1736312288873; numProcessing=1 2025-01-08T05:03:36,045 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/draining/f28094b72178,37583,1736312288873 already deleted, retry=false 2025-01-08T05:03:36,045 INFO [RegionServerTracker-0 {}] master.ServerManager(652): Cluster shutdown set; f28094b72178,37583,1736312288873 expired; onlineServers=2 2025-01-08T05:03:36,045 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [f28094b72178,37967,1736312288941] 2025-01-08T05:03:36,045 DEBUG [RegionServerTracker-0 {}] master.DeadServer(103): Processing f28094b72178,37967,1736312288941; numProcessing=2 2025-01-08T05:03:36,055 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/draining/f28094b72178,37967,1736312288941 already deleted, retry=false 2025-01-08T05:03:36,055 INFO [RegionServerTracker-0 {}] master.ServerManager(652): Cluster shutdown set; f28094b72178,37967,1736312288941 expired; onlineServers=1 2025-01-08T05:03:36,124 DEBUG [RS:0;f28094b72178:39801 {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/oldWALs 2025-01-08T05:03:36,124 INFO [RS:0;f28094b72178:39801 {}] wal.AbstractFSWAL(1074): Closed WAL: AsyncFSWAL f28094b72178%2C39801%2C1736312288779.meta:.meta(num 1736312598109) 2025-01-08T05:03:36,133 WARN [Close-WAL-Writer-0 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(743): complete file /user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/WALs/f28094b72178,39801,1736312288779/f28094b72178%2C39801%2C1736312288779.1736312291555 not finished, retry = 0 2025-01-08T05:03:36,138 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45867 is added to blk_1073741833_1009 (size=6498) 2025-01-08T05:03:36,138 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33991 is added to blk_1073741833_1009 (size=6498) 2025-01-08T05:03:36,138 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40353 is added to blk_1073741833_1009 (size=6498) 2025-01-08T05:03:36,139 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2025-01-08T05:03:36,139 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2025-01-08T05:03:36,139 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37583-0x1000ae1399f0002, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2025-01-08T05:03:36,139 INFO [RS:1;f28094b72178:37583 {}] regionserver.HRegionServer(1307): Exiting; stopping=f28094b72178,37583,1736312288873; zookeeper connection closed. 2025-01-08T05:03:36,140 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37967-0x1000ae1399f0003, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2025-01-08T05:03:36,140 INFO [RS:2;f28094b72178:37967 {}] regionserver.HRegionServer(1307): Exiting; stopping=f28094b72178,37967,1736312288941; zookeeper connection closed. 2025-01-08T05:03:36,140 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@38b772cd {}] hbase.MiniHBaseCluster$SingleFileSystemShutdownThread(216): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@38b772cd 2025-01-08T05:03:36,141 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@205abe83 {}] hbase.MiniHBaseCluster$SingleFileSystemShutdownThread(216): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@205abe83 2025-01-08T05:03:36,235 DEBUG [RS:0;f28094b72178:39801 {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/oldWALs 2025-01-08T05:03:36,236 INFO [RS:0;f28094b72178:39801 {}] wal.AbstractFSWAL(1074): Closed WAL: AsyncFSWAL f28094b72178%2C39801%2C1736312288779:(num 1736312291555) 2025-01-08T05:03:36,236 DEBUG [RS:0;f28094b72178:39801 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:03:36,236 INFO [RS:0;f28094b72178:39801 {}] regionserver.LeaseManager(133): Closed leases 2025-01-08T05:03:36,236 INFO [RS:0;f28094b72178:39801 {}] hbase.ChoreService(370): Chore service for: regionserver/f28094b72178:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS] on shutdown 2025-01-08T05:03:36,236 INFO [regionserver/f28094b72178:0.logRoller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2025-01-08T05:03:36,236 INFO [RS:0;f28094b72178:39801 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.3:39801 2025-01-08T05:03:36,239 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/f28094b72178,39801,1736312288779 2025-01-08T05:03:36,239 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2025-01-08T05:03:36,240 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [f28094b72178,39801,1736312288779] 2025-01-08T05:03:36,240 DEBUG [RegionServerTracker-0 {}] master.DeadServer(103): Processing f28094b72178,39801,1736312288779; numProcessing=3 2025-01-08T05:03:36,242 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/draining/f28094b72178,39801,1736312288779 already deleted, retry=false 2025-01-08T05:03:36,242 INFO [RegionServerTracker-0 {}] master.ServerManager(652): Cluster shutdown set; f28094b72178,39801,1736312288779 expired; onlineServers=0 2025-01-08T05:03:36,242 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2561): ***** STOPPING region server 'f28094b72178,42195,1736312287279' ***** 2025-01-08T05:03:36,242 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2575): STOPPED: Cluster shutdown set; onlineServer=0 2025-01-08T05:03:36,242 DEBUG [M:0;f28094b72178:42195 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@61651d75, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=f28094b72178/172.17.0.3:0 2025-01-08T05:03:36,242 INFO [M:0;f28094b72178:42195 {}] regionserver.HRegionServer(1224): stopping server f28094b72178,42195,1736312287279 2025-01-08T05:03:36,242 INFO [M:0;f28094b72178:42195 {}] regionserver.HRegionServer(1250): stopping server f28094b72178,42195,1736312287279; all regions closed. 2025-01-08T05:03:36,242 DEBUG [M:0;f28094b72178:42195 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T05:03:36,242 DEBUG [M:0;f28094b72178:42195 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2025-01-08T05:03:36,242 DEBUG [M:0;f28094b72178:42195 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2025-01-08T05:03:36,242 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2025-01-08T05:03:36,243 DEBUG [master/f28094b72178:0:becomeActiveMaster-HFileCleaner.large.0-1736312291078 {}] cleaner.HFileCleaner(306): Exit Thread[master/f28094b72178:0:becomeActiveMaster-HFileCleaner.large.0-1736312291078,5,FailOnTimeoutGroup] 2025-01-08T05:03:36,243 INFO [M:0;f28094b72178:42195 {}] hbase.ChoreService(370): Chore service for: master/f28094b72178:0 had [] on shutdown 2025-01-08T05:03:36,243 DEBUG [M:0;f28094b72178:42195 {}] master.HMaster(1744): Stopping service threads 2025-01-08T05:03:36,243 INFO [M:0;f28094b72178:42195 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2025-01-08T05:03:36,243 DEBUG [master/f28094b72178:0:becomeActiveMaster-HFileCleaner.small.0-1736312291094 {}] cleaner.HFileCleaner(306): Exit Thread[master/f28094b72178:0:becomeActiveMaster-HFileCleaner.small.0-1736312291094,5,FailOnTimeoutGroup] 2025-01-08T05:03:36,244 INFO [M:0;f28094b72178:42195 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2025-01-08T05:03:36,244 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2025-01-08T05:03:36,244 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2025-01-08T05:03:36,244 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T05:03:36,244 DEBUG [M:0;f28094b72178:42195 {}] zookeeper.ZKUtil(347): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2025-01-08T05:03:36,244 WARN [M:0;f28094b72178:42195 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2025-01-08T05:03:36,244 INFO [M:0;f28094b72178:42195 {}] assignment.AssignmentManager(391): Stopping assignment manager 2025-01-08T05:03:36,245 INFO [M:0;f28094b72178:42195 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2025-01-08T05:03:36,245 DEBUG [M:0;f28094b72178:42195 {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2025-01-08T05:03:36,246 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2025-01-08T05:03:36,261 INFO [M:0;f28094b72178:42195 {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2025-01-08T05:03:36,261 DEBUG [M:0;f28094b72178:42195 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2025-01-08T05:03:36,261 DEBUG [M:0;f28094b72178:42195 {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2025-01-08T05:03:36,261 DEBUG [M:0;f28094b72178:42195 {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2025-01-08T05:03:36,262 INFO [M:0;f28094b72178:42195 {}] regionserver.HRegion(2837): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=821.01 KB heapSize=985.48 KB 2025-01-08T05:03:36,262 ERROR [AsyncFSWAL-0-hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/MasterData-prefix:f28094b72178,42195,1736312287279 {}] server.NIOServerCnxnFactory(85): Thread Thread[AsyncFSWAL-0-hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/MasterData-prefix:f28094b72178,42195,1736312287279,5,FailOnTimeoutGroup] died java.lang.NullPointerException: Cannot invoke "org.apache.hbase.thirdparty.io.netty.buffer.ByteBuf.readableBytes()" because "this.buf" is null at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutput.buffered(FanOutOneBlockAsyncDFSOutput.java:419) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.wal.AsyncProtobufLogWriter.append(AsyncProtobufLogWriter.java:132) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.doAppend(AsyncFSWAL.java:830) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.doAppend(AsyncFSWAL.java:128) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.appendEntry(AbstractFSWAL.java:1148) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.appendAndSync(AsyncFSWAL.java:500) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.consume(AsyncFSWAL.java:603) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2025-01-08T05:03:36,340 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2025-01-08T05:03:36,340 INFO [RS:0;f28094b72178:39801 {}] regionserver.HRegionServer(1307): Exiting; stopping=f28094b72178,39801,1736312288779; zookeeper connection closed. 2025-01-08T05:03:36,340 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39801-0x1000ae1399f0001, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2025-01-08T05:03:36,341 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@4409693d {}] hbase.MiniHBaseCluster$SingleFileSystemShutdownThread(216): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@4409693d 2025-01-08T05:03:36,341 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2025-01-08T05:03:38,345 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=Coprocessor.RegionServer.CP_org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:03:38,345 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_meta 2025-01-08T05:03:38,346 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_namespace 2025-01-08T05:03:38,346 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_acl 2025-01-08T05:03:38,346 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: Master,sub=Coprocessor.Master.CP_org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:03:38,346 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: Master,sub=Coprocessor.Master.CP_org.apache.hadoop.hbase.security.access.SecureTestUtil$MasterSyncObserver 2025-01-08T05:03:38,346 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T05:03:38,346 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2025-01-08T05:03:38,346 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testExportExpiredSnapshot 2025-01-08T05:03:41,358 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T05:04:05,794 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T05:04:35,795 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. Process Thread Dump: Automatic Stack Trace every 60 seconds waiting on M:0;f28094b72178:42195 226 active threads Thread 1 (main): State: TIMED_WAITING Blocked count: 3 Waited count: 4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.FutureTask.awaitDone(FutureTask.java:444) java.base@17.0.11/java.util.concurrent.FutureTask.get(FutureTask.java:203) app//org.junit.internal.runners.statements.FailOnTimeout.getResult(FailOnTimeout.java:167) app//org.junit.internal.runners.statements.FailOnTimeout.evaluate(FailOnTimeout.java:128) app//org.apache.hadoop.hbase.SystemExitRule$1.evaluate(SystemExitRule.java:39) app//org.junit.rules.RunRules.evaluate(RunRules.java:20) app//org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) app//org.junit.runners.ParentRunner.run(ParentRunner.java:413) app//org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:316) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:240) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:214) app//org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:155) app//org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:385) app//org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) app//org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:507) app//org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:495) Thread 2 (Reference Handler): State: RUNNABLE Blocked count: 9 Waited count: 0 Stack: java.base@17.0.11/java.lang.ref.Reference.waitForReferencePendingList(Native Method) java.base@17.0.11/java.lang.ref.Reference.processPendingReferences(Reference.java:253) java.base@17.0.11/java.lang.ref.Reference$ReferenceHandler.run(Reference.java:215) Thread 3 (Finalizer): State: WAITING Blocked count: 18 Waited count: 11 Waiting on java.lang.ref.ReferenceQueue$Lock@2d1af922 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) java.base@17.0.11/java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:172) Thread 4 (Signal Dispatcher): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 12 (Common-Cleaner): State: TIMED_WAITING Blocked count: 11 Waited count: 17 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/jdk.internal.ref.CleanerImpl.run(CleanerImpl.java:140) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) java.base@17.0.11/jdk.internal.misc.InnocuousThread.run(InnocuousThread.java:162) Thread 13 (Notification Thread): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 14 (pool-1-thread-1): State: RUNNABLE Blocked count: 0 Waited count: 17 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.poll(EPollPort.java:200) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:281) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 15 (pool-1-thread-2): State: WAITING Blocked count: 0 Waited count: 25 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5ff037ac Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:275) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 16 (surefire-forkedjvm-stream-flusher): State: TIMED_WAITING Blocked count: 0 Waited count: 3972 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 18 (surefire-forkedjvm-command-thread): State: WAITING Blocked count: 0 Waited count: 41 Waiting on java.util.concurrent.CountDownLatch$Sync@30623f8a Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.CountDownLatch.await(CountDownLatch.java:230) java.base@17.0.11/sun.nio.ch.PendingFuture.get(PendingFuture.java:178) app//org.apache.maven.surefire.api.util.internal.Channels$2.read(Channels.java:127) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:284) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) app//org.apache.maven.surefire.api.util.internal.Channels$3.readImpl(Channels.java:169) app//org.apache.maven.surefire.api.util.internal.AbstractNoninterruptibleReadableChannel.read(AbstractNoninterruptibleReadableChannel.java:50) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:430) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:419) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.readMessageType(AbstractStreamDecoder.java:116) app//org.apache.maven.surefire.booter.stream.CommandDecoder.decode(CommandDecoder.java:77) app//org.apache.maven.surefire.booter.spi.CommandChannelDecoder.decode(CommandChannelDecoder.java:60) app//org.apache.maven.surefire.booter.CommandReader$CommandRunnable.run(CommandReader.java:290) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 22 (Time-limited test): State: RUNNABLE Blocked count: 11521 Waited count: 12102 Stack: java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo1(Native Method) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:197) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:154) app//org.apache.hadoop.hbase.util.ReflectionUtils.printThreadInfo(ReflectionUtils.java:181) app//org.apache.hadoop.hbase.util.Threads.printThreadInfo(Threads.java:186) app//org.apache.hadoop.hbase.util.Threads.threadDumpingIsAlive(Threads.java:113) app//org.apache.hadoop.hbase.LocalHBaseCluster.join(LocalHBaseCluster.java:394) app//org.apache.hadoop.hbase.MiniHBaseCluster.waitUntilShutDown(MiniHBaseCluster.java:921) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniHBaseCluster(HBaseTestingUtility.java:1359) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniCluster(HBaseTestingUtility.java:1341) app//org.apache.hadoop.hbase.snapshot.TestExportSnapshot.tearDownAfterClass(TestExportSnapshot.java:121) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) java.base@17.0.11/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) java.base@17.0.11/java.lang.reflect.Method.invoke(Method.java:568) app//org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) app//org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) app//org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) app//org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) app//org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) Thread 23 (org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner): State: WAITING Blocked count: 9 Waited count: 10 Waiting on java.lang.ref.ReferenceQueue$Lock@3ae830cc Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) app//org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner.run(FileSystem.java:4171) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 25 (SSL Certificates Store Monitor): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.TaskQueue@647b2d6a Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 34 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@5eeedeb8): State: TIMED_WAITING Blocked count: 0 Waited count: 788 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 35 (GcTimeMonitor obsWindow = 60000, sleepInterval = 5000, maxGcTimePerc = 100): State: TIMED_WAITING Blocked count: 0 Waited count: 79 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.GcTimeMonitor.run(GcTimeMonitor.java:161) Thread 36 (pool-6-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 37 (qtp338304781-37): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 38 (qtp338304781-38): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 39 (qtp338304781-39): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 40 (qtp338304781-40): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 41 (qtp338304781-41-acceptor-0@622697b3-ServerConnector@61d1e83b{HTTP/1.1, (http/1.1)}{localhost:40771}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 42 (qtp338304781-42): State: TIMED_WAITING Blocked count: 0 Waited count: 7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 43 (qtp338304781-43): State: TIMED_WAITING Blocked count: 0 Waited count: 7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 44 (qtp338304781-44): State: TIMED_WAITING Blocked count: 0 Waited count: 7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 45 (Session-HouseKeeper-27700339-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 46 (pool-7-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 53 (FSEditLogAsync): State: WAITING Blocked count: 22 Waited count: 3310 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@75564eb2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.dequeueEdit(FSEditLogAsync.java:241) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.run(FSEditLogAsync.java:250) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 55 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 56 (IPC Server idle connection scanner for port 45939): State: TIMED_WAITING Blocked count: 1 Waited count: 41 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 58 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 79 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 61 (org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor@35667e22): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor.run(PendingReconstructionBlocks.java:267) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 62 (DatanodeAdminMonitor-0): State: TIMED_WAITING Blocked count: 0 Waited count: 131 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 49 (org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor@413afc06): State: TIMED_WAITING Blocked count: 0 Waited count: 79 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor.run(HeartbeatManager.java:563) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 47 (RedundancyMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 132 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) java.base@17.0.11/java.lang.Thread.sleep(Thread.java:344) java.base@17.0.11/java.util.concurrent.TimeUnit.sleep(TimeUnit.java:446) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$RedundancyMonitor.run(BlockManager.java:5352) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 48 (MarkedDeleteBlockScrubberThread): State: TIMED_WAITING Blocked count: 0 Waited count: 38253 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$MarkedDeleteBlockScrubber.run(BlockManager.java:5326) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 51 (Block report processor): State: WAITING Blocked count: 1 Waited count: 1555 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5e061a22 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.processQueue(BlockManager.java:5627) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.run(BlockManager.java:5614) Thread 57 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 54 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 64 (IPC Server handler 0 on default port 45939): State: TIMED_WAITING Blocked count: 63 Waited count: 2336 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 65 (IPC Server handler 1 on default port 45939): State: TIMED_WAITING Blocked count: 74 Waited count: 2356 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 66 (IPC Server handler 2 on default port 45939): State: TIMED_WAITING Blocked count: 63 Waited count: 2336 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 67 (IPC Server handler 3 on default port 45939): State: TIMED_WAITING Blocked count: 90 Waited count: 2350 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 68 (IPC Server handler 4 on default port 45939): State: TIMED_WAITING Blocked count: 66 Waited count: 2345 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 69 (pool-12-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 71 (org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor@32d56002): State: TIMED_WAITING Blocked count: 0 Waited count: 197 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor.run(LeaseManager.java:537) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 72 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor@351f5e8c): State: TIMED_WAITING Blocked count: 0 Waited count: 79 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor.run(FSNamesystem.java:4550) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 73 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller@40208336): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller.run(FSNamesystem.java:4592) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 74 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber@1d776204): State: TIMED_WAITING Blocked count: 0 Waited count: 3 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber.run(FSNamesystem.java:4689) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 75 (CacheReplicationMonitor(1494370843)): State: TIMED_WAITING Blocked count: 0 Waited count: 15 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1759) app//org.apache.hadoop.hdfs.server.blockmanagement.CacheReplicationMonitor.run(CacheReplicationMonitor.java:186) Thread 86 (pool-18-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 87 (qtp12592271-87): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 88 (qtp12592271-88-acceptor-0@1c732bcf-ServerConnector@7455429{HTTP/1.1, (http/1.1)}{localhost:33939}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 89 (qtp12592271-89): State: TIMED_WAITING Blocked count: 0 Waited count: 7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 90 (qtp12592271-90): State: TIMED_WAITING Blocked count: 0 Waited count: 7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 91 (Session-HouseKeeper-6dd22c37-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 92 (nioEventLoopGroup-2-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 93 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@39fef474): State: TIMED_WAITING Blocked count: 0 Waited count: 784 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 95 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 96 (IPC Server idle connection scanner for port 45943): State: TIMED_WAITING Blocked count: 1 Waited count: 41 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 98 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 79 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 101 (Command processor): State: WAITING Blocked count: 2 Waited count: 260 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@29a0b510 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 102 (BP-1446750426-172.17.0.3-1736312281490 heartbeating to localhost/127.0.0.1:45939): State: TIMED_WAITING Blocked count: 1375 Waited count: 1524 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 103 (pool-20-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 85 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@4f29368a): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 97 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 94 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 104 (IPC Server handler 0 on default port 45943): State: TIMED_WAITING Blocked count: 0 Waited count: 394 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 105 (IPC Server handler 1 on default port 45943): State: TIMED_WAITING Blocked count: 0 Waited count: 392 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 106 (IPC Server handler 2 on default port 45943): State: TIMED_WAITING Blocked count: 0 Waited count: 393 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 107 (IPC Server handler 3 on default port 45943): State: TIMED_WAITING Blocked count: 0 Waited count: 393 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 108 (IPC Server handler 4 on default port 45943): State: TIMED_WAITING Blocked count: 0 Waited count: 393 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 120 (pool-26-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 121 (qtp1040803327-121): State: RUNNABLE Blocked count: 1 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 122 (qtp1040803327-122-acceptor-0@7588b8a-ServerConnector@673672ee{HTTP/1.1, (http/1.1)}{localhost:39721}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 118 (IPC Client (366384994) connection to localhost/127.0.0.1:45939 from jenkins): State: TIMED_WAITING Blocked count: 1405 Waited count: 1402 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Thread 123 (qtp1040803327-123): State: TIMED_WAITING Blocked count: 0 Waited count: 7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 124 (qtp1040803327-124): State: TIMED_WAITING Blocked count: 0 Waited count: 7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 119 (IPC Parameter Sending Thread for localhost/127.0.0.1:45939): State: TIMED_WAITING Blocked count: 0 Waited count: 1988 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 125 (Session-HouseKeeper-425819f9-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 126 (nioEventLoopGroup-4-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 127 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@49a955f5): State: TIMED_WAITING Blocked count: 0 Waited count: 784 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 129 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 130 (IPC Server idle connection scanner for port 36851): State: TIMED_WAITING Blocked count: 1 Waited count: 41 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 132 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 79 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 137 (Command processor): State: WAITING Blocked count: 1 Waited count: 249 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@50b25eec Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 138 (BP-1446750426-172.17.0.3-1736312281490 heartbeating to localhost/127.0.0.1:45939): State: TIMED_WAITING Blocked count: 1402 Waited count: 1518 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 139 (pool-29-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 117 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@710dca53): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 131 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 128 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 140 (IPC Server handler 0 on default port 36851): State: TIMED_WAITING Blocked count: 0 Waited count: 392 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 141 (IPC Server handler 1 on default port 36851): State: TIMED_WAITING Blocked count: 0 Waited count: 392 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 142 (IPC Server handler 2 on default port 36851): State: TIMED_WAITING Blocked count: 0 Waited count: 392 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 143 (IPC Server handler 3 on default port 36851): State: TIMED_WAITING Blocked count: 0 Waited count: 392 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 144 (IPC Server handler 4 on default port 36851): State: TIMED_WAITING Blocked count: 0 Waited count: 392 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 154 (pool-36-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 156 (qtp244728962-156): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 157 (qtp244728962-157-acceptor-0@7ea21e77-ServerConnector@22e0a731{HTTP/1.1, (http/1.1)}{localhost:40719}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 158 (qtp244728962-158): State: TIMED_WAITING Blocked count: 0 Waited count: 7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 159 (qtp244728962-159): State: TIMED_WAITING Blocked count: 0 Waited count: 7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 160 (Session-HouseKeeper-562012e4-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 162 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data1)): State: TIMED_WAITING Blocked count: 2 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 164 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data2)): State: TIMED_WAITING Blocked count: 4 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 163 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data3)): State: TIMED_WAITING Blocked count: 25 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 165 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data4)): State: TIMED_WAITING Blocked count: 9 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 173 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data3/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 176 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data1/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 175 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data2/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 177 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data4/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 184 (nioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 188 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@1badcc77): State: TIMED_WAITING Blocked count: 0 Waited count: 783 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 193 (pool-23-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 190 (pool-15-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 192 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 200 (IPC Server idle connection scanner for port 38935): State: TIMED_WAITING Blocked count: 1 Waited count: 41 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 199 (java.util.concurrent.ThreadPoolExecutor$Worker@3412cacb[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 196 (java.util.concurrent.ThreadPoolExecutor$Worker@17da2a7b[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 202 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 79 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 205 (Command processor): State: WAITING Blocked count: 2 Waited count: 290 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7c245a19 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 206 (BP-1446750426-172.17.0.3-1736312281490 heartbeating to localhost/127.0.0.1:45939): State: TIMED_WAITING Blocked count: 1354 Waited count: 1514 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 207 (pool-46-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 153 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@4ee4c44b): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 201 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 191 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 208 (IPC Server handler 0 on default port 38935): State: TIMED_WAITING Blocked count: 0 Waited count: 394 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 209 (IPC Server handler 1 on default port 38935): State: TIMED_WAITING Blocked count: 0 Waited count: 392 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 210 (IPC Server handler 2 on default port 38935): State: TIMED_WAITING Blocked count: 0 Waited count: 392 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 211 (IPC Server handler 3 on default port 38935): State: TIMED_WAITING Blocked count: 0 Waited count: 392 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 213 (IPC Server handler 4 on default port 38935): State: TIMED_WAITING Blocked count: 0 Waited count: 392 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 215 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data5)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 216 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data6)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 220 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data5/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 222 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data6/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 226 (pool-33-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 229 (java.util.concurrent.ThreadPoolExecutor$Worker@3f6c1323[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 234 (FsDatasetAsyncDiskServiceFixer): State: TIMED_WAITING Blocked count: 0 Waited count: 14 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer.run(HBaseTestingUtility.java:599) Thread 236 (NIOServerCxnFactory.SelectorThread-0): State: RUNNABLE Blocked count: 4 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 237 (NIOServerCxnFactory.SelectorThread-1): State: RUNNABLE Blocked count: 10 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 238 (NIOServerCxnFactory.AcceptThread:localhost/127.0.0.1:59048): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.select(NIOServerCnxnFactory.java:205) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.run(NIOServerCnxnFactory.java:181) Thread 235 (ConnnectionExpirer): State: TIMED_WAITING Blocked count: 0 Waited count: 40 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.NIOServerCnxnFactory$ConnectionExpirerThread.run(NIOServerCnxnFactory.java:554) Thread 239 (SessionTracker): State: TIMED_WAITING Blocked count: 0 Waited count: 196 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.SessionTrackerImpl.run(SessionTrackerImpl.java:163) Thread 240 (SyncThread:0): State: WAITING Blocked count: 22 Waited count: 754 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@51109921 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.SyncRequestProcessor.run(SyncRequestProcessor.java:170) Thread 241 (ProcessThread(sid:0 cport:59048):): State: WAITING Blocked count: 1 Waited count: 858 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5a02804a Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.PrepRequestProcessor.run(PrepRequestProcessor.java:142) Thread 242 (RequestThrottler): State: WAITING Blocked count: 0 Waited count: 882 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5cd4ed15 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.RequestThrottler.run(RequestThrottler.java:147) Thread 243 (NIOWorkerThread-1): State: WAITING Blocked count: 3 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 245 (LeaseRenewer:jenkins@localhost:45939): State: TIMED_WAITING Blocked count: 10 Waited count: 403 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 254 (Time-limited test.named-queue-events-pool-0): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@543bccbd Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) app//com.lmax.disruptor.BlockingWaitStrategy.waitFor(BlockingWaitStrategy.java:47) app//com.lmax.disruptor.ProcessingSequenceBarrier.waitFor(ProcessingSequenceBarrier.java:56) app//com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:159) app//com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 255 (HBase-Metrics2-1): State: TIMED_WAITING Blocked count: 0 Waited count: 325 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 256 (RS-EventLoopGroup-1-1): State: RUNNABLE Blocked count: 21 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 257 (Time-limited test-SendThread(127.0.0.1:59048)): State: RUNNABLE Blocked count: 42 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:332) app//org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1289) Thread 258 (Time-limited test-EventThread): State: WAITING Blocked count: 1 Waited count: 59 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@61b2fcae Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:550) Thread 259 (NIOWorkerThread-2): State: WAITING Blocked count: 2 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 260 (NIOWorkerThread-3): State: WAITING Blocked count: 2 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 261 (NIOWorkerThread-4): State: WAITING Blocked count: 1 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 262 (zk-event-processor-pool-0): State: WAITING Blocked count: 38 Waited count: 91 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5c993aff Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 263 (NIOWorkerThread-5): State: WAITING Blocked count: 3 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 264 (NIOWorkerThread-6): State: WAITING Blocked count: 2 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 265 (NIOWorkerThread-7): State: WAITING Blocked count: 3 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 266 (NIOWorkerThread-8): State: WAITING Blocked count: 1 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 267 (NIOWorkerThread-9): State: WAITING Blocked count: 3 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 268 (NIOWorkerThread-10): State: WAITING Blocked count: 2 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 269 (NIOWorkerThread-11): State: WAITING Blocked count: 1 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 270 (NIOWorkerThread-12): State: WAITING Blocked count: 0 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 271 (NIOWorkerThread-13): State: WAITING Blocked count: 2 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 272 (NIOWorkerThread-14): State: WAITING Blocked count: 2 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 273 (NIOWorkerThread-15): State: WAITING Blocked count: 4 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 274 (NIOWorkerThread-16): State: WAITING Blocked count: 2 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 275 (RpcServer.default.FPBQ.Fifo.handler=0,queue=0,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@564d5bf5 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 276 (RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195): State: WAITING Blocked count: 109 Waited count: 415 Waiting on java.util.concurrent.Semaphore$NonfairSync@643f078a Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 277 (RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195): State: WAITING Blocked count: 203 Waited count: 823 Waiting on java.util.concurrent.Semaphore$NonfairSync@129ed02f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 278 (RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42195): State: WAITING Blocked count: 38 Waited count: 9161 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7c54e41 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 279 (RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@3b6b3fbf Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 280 (RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@3b6b3fbf Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 281 (RpcServer.replication.FPBQ.Fifo.handler=0,queue=0,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@27f47c0f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 282 (RpcServer.replication.FPBQ.Fifo.handler=1,queue=0,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@438fcf76 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 283 (RpcServer.replication.FPBQ.Fifo.handler=2,queue=0,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@6e953600 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 284 (RpcServer.metaPriority.FPBQ.Fifo.handler=0,queue=0,port=42195): State: WAITING Blocked count: 0 Waited count: 7 Waiting on java.util.concurrent.Semaphore$NonfairSync@ce17aaa Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 288 (RS-EventLoopGroup-3-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 308 (RS-EventLoopGroup-4-1): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 328 (RS-EventLoopGroup-5-1): State: RUNNABLE Blocked count: 68 Waited count: 3 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 286 (M:0;f28094b72178:42195): State: TIMED_WAITING Blocked count: 8 Waited count: 3268 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1759) app//org.apache.hadoop.hbase.regionserver.wal.SyncFuture.get(SyncFuture.java:169) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.blockOnSync(AbstractFSWAL.java:879) app//org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.doSync(AsyncFSWAL.java:671) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.lambda$sync$1(AbstractFSWAL.java:600) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL$$Lambda$957/0x00007ffa3cf1b800.run(Unknown Source) app//org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:600) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:590) app//org.apache.hadoop.hbase.regionserver.HRegion.doSyncOfUnflushedWALChanges(HRegion.java:2869) app//org.apache.hadoop.hbase.regionserver.HRegion.internalPrepareFlushCache(HRegion.java:2811) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2670) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2644) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2635) app//org.apache.hadoop.hbase.regionserver.HRegion.doClose(HRegion.java:1810) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1631) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1586) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1569) Thread 349 (Monitor thread for TaskMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 39 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.monitoring.TaskMonitor$MonitorRunnable.run(TaskMonitor.java:325) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 351 (master/f28094b72178:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 353 (master/f28094b72178:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 355 (org.apache.hadoop.hdfs.PeerCache@7594f538): State: TIMED_WAITING Blocked count: 0 Waited count: 130 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.PeerCache.run(PeerCache.java:253) app//org.apache.hadoop.hdfs.PeerCache.access$000(PeerCache.java:46) app//org.apache.hadoop.hdfs.PeerCache$1.run(PeerCache.java:124) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 374 (master:store-WAL-Roller): State: TIMED_WAITING Blocked count: 0 Waited count: 3855 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hbase.wal.AbstractWALRoller.run(AbstractWALRoller.java:179) Thread 391 (RS-EventLoopGroup-5-2): State: RUNNABLE Blocked count: 53 Waited count: 4 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 392 (RS-EventLoopGroup-5-3): State: RUNNABLE Blocked count: 37 Waited count: 4 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 404 (Idle-Rpc-Conn-Sweeper-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 62 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 415 (SnapshotHandlerChoreCleaner): State: TIMED_WAITING Blocked count: 0 Waited count: 39 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 403 (RpcClient-timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 38495 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 424 (RS-EventLoopGroup-1-2): State: RUNNABLE Blocked count: 50 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 425 (RS-EventLoopGroup-1-3): State: RUNNABLE Blocked count: 34 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 440 (RegionServerTracker-0): State: WAITING Blocked count: 8 Waited count: 11 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4814fa0 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 471 (regionserver/f28094b72178:0.procedureResultReporter): State: WAITING Blocked count: 13 Waited count: 27 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@21db0ca7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 476 (regionserver/f28094b72178:0.procedureResultReporter): State: WAITING Blocked count: 19 Waited count: 39 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b2ba608 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 467 (regionserver/f28094b72178:0.procedureResultReporter): State: WAITING Blocked count: 9 Waited count: 19 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7f063557 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 493 (LeaseRenewer:jenkins.hfs.0@localhost:45939): State: TIMED_WAITING Blocked count: 10 Waited count: 401 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 507 (LeaseRenewer:jenkins.hfs.2@localhost:45939): State: TIMED_WAITING Blocked count: 10 Waited count: 400 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 508 (LeaseRenewer:jenkins.hfs.1@localhost:45939): State: TIMED_WAITING Blocked count: 10 Waited count: 401 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 512 (RS-EventLoopGroup-4-2): State: RUNNABLE Blocked count: 2 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 523 (RS-EventLoopGroup-4-3): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 525 (region-location-0): State: WAITING Blocked count: 7 Waited count: 12 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@71749958 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 551 (Async-Client-Retry-Timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 38228 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 565 (RPCClient-NioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 8 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 582 (ForkJoinPool.commonPool-worker-1): State: WAITING Blocked count: 0 Waited count: 380 Waiting on java.util.concurrent.ForkJoinPool@8b31de5 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1724) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Thread 583 (ForkJoinPool.commonPool-worker-2): State: WAITING Blocked count: 0 Waited count: 43 Waiting on java.util.concurrent.ForkJoinPool@8b31de5 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1724) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Thread 604 (region-location-1): State: WAITING Blocked count: 6 Waited count: 11 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@71749958 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 605 (region-location-2): State: WAITING Blocked count: 4 Waited count: 9 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@71749958 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 606 (region-location-3): State: WAITING Blocked count: 3 Waited count: 8 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@71749958 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1007 (MutableQuantiles-0): State: TIMED_WAITING Blocked count: 0 Waited count: 414 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1071 (RPCClient-NioEventLoopGroup-6-2): State: RUNNABLE Blocked count: 26 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1111 (zk-permission-watcher-pool-0): State: WAITING Blocked count: 61 Waited count: 92 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@8cc91a Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1119 (RPCClient-NioEventLoopGroup-6-3): State: RUNNABLE Blocked count: 5 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1567 (Container metrics unregistration): State: WAITING Blocked count: 10 Waited count: 33 Waiting on java.util.TaskQueue@55e73e0a Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 1742 (ForkJoinPool.commonPool-worker-4): State: WAITING Blocked count: 0 Waited count: 584 Waiting on java.util.concurrent.ForkJoinPool@8b31de5 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1724) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Thread 1743 (ForkJoinPool.commonPool-worker-5): State: TIMED_WAITING Blocked count: 0 Waited count: 465 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkUntil(LockSupport.java:410) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1726) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Thread 2046 (RS-EventLoopGroup-3-2): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 2058 (RPCClient-NioEventLoopGroup-6-4): State: RUNNABLE Blocked count: 2 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 2059 (RS-EventLoopGroup-3-3): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 2420 (region-location-4): State: WAITING Blocked count: 2 Waited count: 7 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@71749958 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5232 (RPCClient-NioEventLoopGroup-6-5): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5233 (RPCClient-NioEventLoopGroup-6-6): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5234 (RPCClient-NioEventLoopGroup-6-7): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9459 (AsyncFSWAL-1-hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/MasterData-prefix:f28094b72178,42195,1736312287279): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@452b5d66 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9462 (Timer for 'JobHistoryServer' metrics system): State: TIMED_WAITING Blocked count: 0 Waited count: 6 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) 2025-01-08T05:05:05,795 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T05:05:35,795 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. Process Thread Dump: Automatic Stack Trace every 60 seconds waiting on M:0;f28094b72178:42195 221 active threads Thread 1 (main): State: TIMED_WAITING Blocked count: 3 Waited count: 4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.FutureTask.awaitDone(FutureTask.java:444) java.base@17.0.11/java.util.concurrent.FutureTask.get(FutureTask.java:203) app//org.junit.internal.runners.statements.FailOnTimeout.getResult(FailOnTimeout.java:167) app//org.junit.internal.runners.statements.FailOnTimeout.evaluate(FailOnTimeout.java:128) app//org.apache.hadoop.hbase.SystemExitRule$1.evaluate(SystemExitRule.java:39) app//org.junit.rules.RunRules.evaluate(RunRules.java:20) app//org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) app//org.junit.runners.ParentRunner.run(ParentRunner.java:413) app//org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:316) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:240) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:214) app//org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:155) app//org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:385) app//org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) app//org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:507) app//org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:495) Thread 2 (Reference Handler): State: RUNNABLE Blocked count: 9 Waited count: 0 Stack: java.base@17.0.11/java.lang.ref.Reference.waitForReferencePendingList(Native Method) java.base@17.0.11/java.lang.ref.Reference.processPendingReferences(Reference.java:253) java.base@17.0.11/java.lang.ref.Reference$ReferenceHandler.run(Reference.java:215) Thread 3 (Finalizer): State: WAITING Blocked count: 18 Waited count: 11 Waiting on java.lang.ref.ReferenceQueue$Lock@2d1af922 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) java.base@17.0.11/java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:172) Thread 4 (Signal Dispatcher): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 12 (Common-Cleaner): State: TIMED_WAITING Blocked count: 11 Waited count: 18 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/jdk.internal.ref.CleanerImpl.run(CleanerImpl.java:140) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) java.base@17.0.11/jdk.internal.misc.InnocuousThread.run(InnocuousThread.java:162) Thread 13 (Notification Thread): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 14 (pool-1-thread-1): State: RUNNABLE Blocked count: 0 Waited count: 20 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.poll(EPollPort.java:200) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:281) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 15 (pool-1-thread-2): State: WAITING Blocked count: 0 Waited count: 28 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5ff037ac Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:275) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 16 (surefire-forkedjvm-stream-flusher): State: TIMED_WAITING Blocked count: 0 Waited count: 4572 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 18 (surefire-forkedjvm-command-thread): State: WAITING Blocked count: 0 Waited count: 47 Waiting on java.util.concurrent.CountDownLatch$Sync@1ad6684a Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.CountDownLatch.await(CountDownLatch.java:230) java.base@17.0.11/sun.nio.ch.PendingFuture.get(PendingFuture.java:178) app//org.apache.maven.surefire.api.util.internal.Channels$2.read(Channels.java:127) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:284) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) app//org.apache.maven.surefire.api.util.internal.Channels$3.readImpl(Channels.java:169) app//org.apache.maven.surefire.api.util.internal.AbstractNoninterruptibleReadableChannel.read(AbstractNoninterruptibleReadableChannel.java:50) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:430) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:419) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.readMessageType(AbstractStreamDecoder.java:116) app//org.apache.maven.surefire.booter.stream.CommandDecoder.decode(CommandDecoder.java:77) app//org.apache.maven.surefire.booter.spi.CommandChannelDecoder.decode(CommandChannelDecoder.java:60) app//org.apache.maven.surefire.booter.CommandReader$CommandRunnable.run(CommandReader.java:290) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 22 (Time-limited test): State: RUNNABLE Blocked count: 11521 Waited count: 12103 Stack: java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo1(Native Method) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:197) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:154) app//org.apache.hadoop.hbase.util.ReflectionUtils.printThreadInfo(ReflectionUtils.java:181) app//org.apache.hadoop.hbase.util.Threads.printThreadInfo(Threads.java:186) app//org.apache.hadoop.hbase.util.Threads.threadDumpingIsAlive(Threads.java:113) app//org.apache.hadoop.hbase.LocalHBaseCluster.join(LocalHBaseCluster.java:394) app//org.apache.hadoop.hbase.MiniHBaseCluster.waitUntilShutDown(MiniHBaseCluster.java:921) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniHBaseCluster(HBaseTestingUtility.java:1359) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniCluster(HBaseTestingUtility.java:1341) app//org.apache.hadoop.hbase.snapshot.TestExportSnapshot.tearDownAfterClass(TestExportSnapshot.java:121) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) java.base@17.0.11/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) java.base@17.0.11/java.lang.reflect.Method.invoke(Method.java:568) app//org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) app//org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) app//org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) app//org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) app//org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) Thread 23 (org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner): State: WAITING Blocked count: 9 Waited count: 10 Waiting on java.lang.ref.ReferenceQueue$Lock@3ae830cc Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) app//org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner.run(FileSystem.java:4171) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 25 (SSL Certificates Store Monitor): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.TaskQueue@647b2d6a Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 34 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@5eeedeb8): State: TIMED_WAITING Blocked count: 0 Waited count: 908 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 35 (GcTimeMonitor obsWindow = 60000, sleepInterval = 5000, maxGcTimePerc = 100): State: TIMED_WAITING Blocked count: 0 Waited count: 91 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.GcTimeMonitor.run(GcTimeMonitor.java:161) Thread 36 (pool-6-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 37 (qtp338304781-37): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 38 (qtp338304781-38): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 39 (qtp338304781-39): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 40 (qtp338304781-40): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 41 (qtp338304781-41-acceptor-0@622697b3-ServerConnector@61d1e83b{HTTP/1.1, (http/1.1)}{localhost:40771}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 42 (qtp338304781-42): State: TIMED_WAITING Blocked count: 0 Waited count: 8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 43 (qtp338304781-43): State: TIMED_WAITING Blocked count: 0 Waited count: 8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 44 (qtp338304781-44): State: TIMED_WAITING Blocked count: 0 Waited count: 8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 45 (Session-HouseKeeper-27700339-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 46 (pool-7-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 53 (FSEditLogAsync): State: WAITING Blocked count: 22 Waited count: 3310 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@75564eb2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.dequeueEdit(FSEditLogAsync.java:241) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.run(FSEditLogAsync.java:250) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 55 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 56 (IPC Server idle connection scanner for port 45939): State: TIMED_WAITING Blocked count: 1 Waited count: 47 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 58 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 91 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 61 (org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor@35667e22): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor.run(PendingReconstructionBlocks.java:267) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 62 (DatanodeAdminMonitor-0): State: TIMED_WAITING Blocked count: 0 Waited count: 151 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 49 (org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor@413afc06): State: TIMED_WAITING Blocked count: 0 Waited count: 91 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor.run(HeartbeatManager.java:563) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 47 (RedundancyMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 152 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) java.base@17.0.11/java.lang.Thread.sleep(Thread.java:344) java.base@17.0.11/java.util.concurrent.TimeUnit.sleep(TimeUnit.java:446) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$RedundancyMonitor.run(BlockManager.java:5352) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 48 (MarkedDeleteBlockScrubberThread): State: TIMED_WAITING Blocked count: 0 Waited count: 44215 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$MarkedDeleteBlockScrubber.run(BlockManager.java:5326) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 51 (Block report processor): State: WAITING Blocked count: 1 Waited count: 1555 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5e061a22 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.processQueue(BlockManager.java:5627) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.run(BlockManager.java:5614) Thread 57 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 54 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 64 (IPC Server handler 0 on default port 45939): State: TIMED_WAITING Blocked count: 67 Waited count: 2397 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 65 (IPC Server handler 1 on default port 45939): State: TIMED_WAITING Blocked count: 76 Waited count: 2416 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 66 (IPC Server handler 2 on default port 45939): State: TIMED_WAITING Blocked count: 66 Waited count: 2397 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 67 (IPC Server handler 3 on default port 45939): State: TIMED_WAITING Blocked count: 96 Waited count: 2411 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 68 (IPC Server handler 4 on default port 45939): State: TIMED_WAITING Blocked count: 71 Waited count: 2406 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 69 (pool-12-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 71 (org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor@32d56002): State: TIMED_WAITING Blocked count: 0 Waited count: 227 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor.run(LeaseManager.java:537) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 72 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor@351f5e8c): State: TIMED_WAITING Blocked count: 0 Waited count: 91 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor.run(FSNamesystem.java:4550) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 73 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller@40208336): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller.run(FSNamesystem.java:4592) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 74 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber@1d776204): State: TIMED_WAITING Blocked count: 0 Waited count: 3 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber.run(FSNamesystem.java:4689) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 75 (CacheReplicationMonitor(1494370843)): State: TIMED_WAITING Blocked count: 0 Waited count: 17 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1759) app//org.apache.hadoop.hdfs.server.blockmanagement.CacheReplicationMonitor.run(CacheReplicationMonitor.java:186) Thread 86 (pool-18-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 87 (qtp12592271-87): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 88 (qtp12592271-88-acceptor-0@1c732bcf-ServerConnector@7455429{HTTP/1.1, (http/1.1)}{localhost:33939}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 89 (qtp12592271-89): State: TIMED_WAITING Blocked count: 0 Waited count: 8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 90 (qtp12592271-90): State: TIMED_WAITING Blocked count: 0 Waited count: 8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 91 (Session-HouseKeeper-6dd22c37-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 92 (nioEventLoopGroup-2-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 93 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@39fef474): State: TIMED_WAITING Blocked count: 0 Waited count: 904 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 95 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 96 (IPC Server idle connection scanner for port 45943): State: TIMED_WAITING Blocked count: 1 Waited count: 47 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 98 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 91 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 101 (Command processor): State: WAITING Blocked count: 2 Waited count: 280 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@29a0b510 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 102 (BP-1446750426-172.17.0.3-1736312281490 heartbeating to localhost/127.0.0.1:45939): State: TIMED_WAITING Blocked count: 1397 Waited count: 1578 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 103 (pool-20-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 85 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@4f29368a): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 97 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 94 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 104 (IPC Server handler 0 on default port 45943): State: TIMED_WAITING Blocked count: 0 Waited count: 454 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 105 (IPC Server handler 1 on default port 45943): State: TIMED_WAITING Blocked count: 0 Waited count: 452 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 106 (IPC Server handler 2 on default port 45943): State: TIMED_WAITING Blocked count: 0 Waited count: 453 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 107 (IPC Server handler 3 on default port 45943): State: TIMED_WAITING Blocked count: 0 Waited count: 453 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 108 (IPC Server handler 4 on default port 45943): State: TIMED_WAITING Blocked count: 0 Waited count: 453 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 120 (pool-26-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 121 (qtp1040803327-121): State: RUNNABLE Blocked count: 1 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 122 (qtp1040803327-122-acceptor-0@7588b8a-ServerConnector@673672ee{HTTP/1.1, (http/1.1)}{localhost:39721}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 118 (IPC Client (366384994) connection to localhost/127.0.0.1:45939 from jenkins): State: TIMED_WAITING Blocked count: 1445 Waited count: 1442 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Thread 123 (qtp1040803327-123): State: TIMED_WAITING Blocked count: 0 Waited count: 8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 124 (qtp1040803327-124): State: TIMED_WAITING Blocked count: 0 Waited count: 8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 119 (IPC Parameter Sending Thread for localhost/127.0.0.1:45939): State: TIMED_WAITING Blocked count: 0 Waited count: 2028 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 125 (Session-HouseKeeper-425819f9-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 126 (nioEventLoopGroup-4-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 127 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@49a955f5): State: TIMED_WAITING Blocked count: 0 Waited count: 904 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 129 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 130 (IPC Server idle connection scanner for port 36851): State: TIMED_WAITING Blocked count: 1 Waited count: 47 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 132 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 91 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 137 (Command processor): State: WAITING Blocked count: 1 Waited count: 269 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@50b25eec Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 138 (BP-1446750426-172.17.0.3-1736312281490 heartbeating to localhost/127.0.0.1:45939): State: TIMED_WAITING Blocked count: 1422 Waited count: 1564 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 139 (pool-29-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 117 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@710dca53): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 131 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 128 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 140 (IPC Server handler 0 on default port 36851): State: TIMED_WAITING Blocked count: 0 Waited count: 452 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 141 (IPC Server handler 1 on default port 36851): State: TIMED_WAITING Blocked count: 0 Waited count: 452 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 142 (IPC Server handler 2 on default port 36851): State: TIMED_WAITING Blocked count: 0 Waited count: 452 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 143 (IPC Server handler 3 on default port 36851): State: TIMED_WAITING Blocked count: 0 Waited count: 452 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 144 (IPC Server handler 4 on default port 36851): State: TIMED_WAITING Blocked count: 0 Waited count: 452 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 154 (pool-36-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 156 (qtp244728962-156): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 157 (qtp244728962-157-acceptor-0@7ea21e77-ServerConnector@22e0a731{HTTP/1.1, (http/1.1)}{localhost:40719}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 158 (qtp244728962-158): State: TIMED_WAITING Blocked count: 0 Waited count: 8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 159 (qtp244728962-159): State: TIMED_WAITING Blocked count: 0 Waited count: 8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 160 (Session-HouseKeeper-562012e4-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 162 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data1)): State: TIMED_WAITING Blocked count: 2 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 164 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data2)): State: TIMED_WAITING Blocked count: 4 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 163 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data3)): State: TIMED_WAITING Blocked count: 25 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 165 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data4)): State: TIMED_WAITING Blocked count: 9 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 173 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data3/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 176 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data1/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 175 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data2/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 177 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data4/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 184 (nioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 188 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@1badcc77): State: TIMED_WAITING Blocked count: 0 Waited count: 903 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 193 (pool-23-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 190 (pool-15-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 192 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 200 (IPC Server idle connection scanner for port 38935): State: TIMED_WAITING Blocked count: 1 Waited count: 47 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 199 (java.util.concurrent.ThreadPoolExecutor$Worker@3412cacb[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 196 (java.util.concurrent.ThreadPoolExecutor$Worker@17da2a7b[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 202 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 91 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 205 (Command processor): State: WAITING Blocked count: 2 Waited count: 310 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7c245a19 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 206 (BP-1446750426-172.17.0.3-1736312281490 heartbeating to localhost/127.0.0.1:45939): State: TIMED_WAITING Blocked count: 1374 Waited count: 1554 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 207 (pool-46-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 153 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@4ee4c44b): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 201 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 191 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 208 (IPC Server handler 0 on default port 38935): State: TIMED_WAITING Blocked count: 0 Waited count: 454 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 209 (IPC Server handler 1 on default port 38935): State: TIMED_WAITING Blocked count: 0 Waited count: 452 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 210 (IPC Server handler 2 on default port 38935): State: TIMED_WAITING Blocked count: 0 Waited count: 452 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 211 (IPC Server handler 3 on default port 38935): State: TIMED_WAITING Blocked count: 0 Waited count: 452 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 213 (IPC Server handler 4 on default port 38935): State: TIMED_WAITING Blocked count: 0 Waited count: 452 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 215 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data5)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 216 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data6)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 220 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data5/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 222 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data6/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 226 (pool-33-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 229 (java.util.concurrent.ThreadPoolExecutor$Worker@3f6c1323[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 234 (FsDatasetAsyncDiskServiceFixer): State: TIMED_WAITING Blocked count: 0 Waited count: 16 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer.run(HBaseTestingUtility.java:599) Thread 236 (NIOServerCxnFactory.SelectorThread-0): State: RUNNABLE Blocked count: 4 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 237 (NIOServerCxnFactory.SelectorThread-1): State: RUNNABLE Blocked count: 10 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 238 (NIOServerCxnFactory.AcceptThread:localhost/127.0.0.1:59048): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.select(NIOServerCnxnFactory.java:205) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.run(NIOServerCnxnFactory.java:181) Thread 235 (ConnnectionExpirer): State: TIMED_WAITING Blocked count: 0 Waited count: 46 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.NIOServerCnxnFactory$ConnectionExpirerThread.run(NIOServerCnxnFactory.java:554) Thread 239 (SessionTracker): State: TIMED_WAITING Blocked count: 0 Waited count: 226 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.SessionTrackerImpl.run(SessionTrackerImpl.java:163) Thread 240 (SyncThread:0): State: WAITING Blocked count: 22 Waited count: 759 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@51109921 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.SyncRequestProcessor.run(SyncRequestProcessor.java:170) Thread 241 (ProcessThread(sid:0 cport:59048):): State: WAITING Blocked count: 1 Waited count: 863 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5a02804a Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.PrepRequestProcessor.run(PrepRequestProcessor.java:142) Thread 242 (RequestThrottler): State: WAITING Blocked count: 0 Waited count: 887 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5cd4ed15 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.RequestThrottler.run(RequestThrottler.java:147) Thread 243 (NIOWorkerThread-1): State: WAITING Blocked count: 3 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 254 (Time-limited test.named-queue-events-pool-0): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@543bccbd Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) app//com.lmax.disruptor.BlockingWaitStrategy.waitFor(BlockingWaitStrategy.java:47) app//com.lmax.disruptor.ProcessingSequenceBarrier.waitFor(ProcessingSequenceBarrier.java:56) app//com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:159) app//com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 255 (HBase-Metrics2-1): State: TIMED_WAITING Blocked count: 0 Waited count: 361 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 256 (RS-EventLoopGroup-1-1): State: RUNNABLE Blocked count: 21 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 257 (Time-limited test-SendThread(127.0.0.1:59048)): State: RUNNABLE Blocked count: 42 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:332) app//org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1289) Thread 258 (Time-limited test-EventThread): State: WAITING Blocked count: 1 Waited count: 59 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@61b2fcae Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:550) Thread 259 (NIOWorkerThread-2): State: WAITING Blocked count: 2 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 260 (NIOWorkerThread-3): State: WAITING Blocked count: 2 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 261 (NIOWorkerThread-4): State: WAITING Blocked count: 1 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 262 (zk-event-processor-pool-0): State: WAITING Blocked count: 38 Waited count: 91 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5c993aff Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 263 (NIOWorkerThread-5): State: WAITING Blocked count: 3 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 264 (NIOWorkerThread-6): State: WAITING Blocked count: 2 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 265 (NIOWorkerThread-7): State: WAITING Blocked count: 3 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 266 (NIOWorkerThread-8): State: WAITING Blocked count: 1 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 267 (NIOWorkerThread-9): State: WAITING Blocked count: 3 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 268 (NIOWorkerThread-10): State: WAITING Blocked count: 2 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 269 (NIOWorkerThread-11): State: WAITING Blocked count: 1 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 270 (NIOWorkerThread-12): State: WAITING Blocked count: 0 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 271 (NIOWorkerThread-13): State: WAITING Blocked count: 2 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 272 (NIOWorkerThread-14): State: WAITING Blocked count: 2 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 273 (NIOWorkerThread-15): State: WAITING Blocked count: 4 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 274 (NIOWorkerThread-16): State: WAITING Blocked count: 2 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 275 (RpcServer.default.FPBQ.Fifo.handler=0,queue=0,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@564d5bf5 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 276 (RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195): State: WAITING Blocked count: 109 Waited count: 415 Waiting on java.util.concurrent.Semaphore$NonfairSync@643f078a Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 277 (RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195): State: WAITING Blocked count: 203 Waited count: 823 Waiting on java.util.concurrent.Semaphore$NonfairSync@129ed02f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 278 (RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42195): State: WAITING Blocked count: 38 Waited count: 9161 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7c54e41 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 279 (RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@3b6b3fbf Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 280 (RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@3b6b3fbf Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 281 (RpcServer.replication.FPBQ.Fifo.handler=0,queue=0,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@27f47c0f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 282 (RpcServer.replication.FPBQ.Fifo.handler=1,queue=0,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@438fcf76 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 283 (RpcServer.replication.FPBQ.Fifo.handler=2,queue=0,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@6e953600 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 284 (RpcServer.metaPriority.FPBQ.Fifo.handler=0,queue=0,port=42195): State: WAITING Blocked count: 0 Waited count: 7 Waiting on java.util.concurrent.Semaphore$NonfairSync@ce17aaa Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 288 (RS-EventLoopGroup-3-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 308 (RS-EventLoopGroup-4-1): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 328 (RS-EventLoopGroup-5-1): State: RUNNABLE Blocked count: 68 Waited count: 3 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 286 (M:0;f28094b72178:42195): State: TIMED_WAITING Blocked count: 8 Waited count: 3268 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1759) app//org.apache.hadoop.hbase.regionserver.wal.SyncFuture.get(SyncFuture.java:169) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.blockOnSync(AbstractFSWAL.java:879) app//org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.doSync(AsyncFSWAL.java:671) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.lambda$sync$1(AbstractFSWAL.java:600) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL$$Lambda$957/0x00007ffa3cf1b800.run(Unknown Source) app//org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:600) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:590) app//org.apache.hadoop.hbase.regionserver.HRegion.doSyncOfUnflushedWALChanges(HRegion.java:2869) app//org.apache.hadoop.hbase.regionserver.HRegion.internalPrepareFlushCache(HRegion.java:2811) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2670) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2644) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2635) app//org.apache.hadoop.hbase.regionserver.HRegion.doClose(HRegion.java:1810) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1631) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1586) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1569) Thread 349 (Monitor thread for TaskMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 45 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.monitoring.TaskMonitor$MonitorRunnable.run(TaskMonitor.java:325) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 351 (master/f28094b72178:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 353 (master/f28094b72178:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 355 (org.apache.hadoop.hdfs.PeerCache@7594f538): State: TIMED_WAITING Blocked count: 0 Waited count: 150 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.PeerCache.run(PeerCache.java:253) app//org.apache.hadoop.hdfs.PeerCache.access$000(PeerCache.java:46) app//org.apache.hadoop.hdfs.PeerCache$1.run(PeerCache.java:124) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 374 (master:store-WAL-Roller): State: TIMED_WAITING Blocked count: 0 Waited count: 4455 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hbase.wal.AbstractWALRoller.run(AbstractWALRoller.java:179) Thread 391 (RS-EventLoopGroup-5-2): State: RUNNABLE Blocked count: 53 Waited count: 4 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 392 (RS-EventLoopGroup-5-3): State: RUNNABLE Blocked count: 37 Waited count: 4 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 404 (Idle-Rpc-Conn-Sweeper-pool-0): State: WAITING Blocked count: 0 Waited count: 67 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@75e9f460 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 415 (SnapshotHandlerChoreCleaner): State: TIMED_WAITING Blocked count: 0 Waited count: 45 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 403 (RpcClient-timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 44495 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 424 (RS-EventLoopGroup-1-2): State: RUNNABLE Blocked count: 50 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 425 (RS-EventLoopGroup-1-3): State: RUNNABLE Blocked count: 34 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 440 (RegionServerTracker-0): State: WAITING Blocked count: 8 Waited count: 11 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4814fa0 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 471 (regionserver/f28094b72178:0.procedureResultReporter): State: WAITING Blocked count: 13 Waited count: 27 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@21db0ca7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 476 (regionserver/f28094b72178:0.procedureResultReporter): State: WAITING Blocked count: 19 Waited count: 39 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b2ba608 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 467 (regionserver/f28094b72178:0.procedureResultReporter): State: WAITING Blocked count: 9 Waited count: 19 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7f063557 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 512 (RS-EventLoopGroup-4-2): State: RUNNABLE Blocked count: 2 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 523 (RS-EventLoopGroup-4-3): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 525 (region-location-0): State: WAITING Blocked count: 7 Waited count: 12 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@71749958 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 551 (Async-Client-Retry-Timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 44230 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 565 (RPCClient-NioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 8 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 582 (ForkJoinPool.commonPool-worker-1): State: TIMED_WAITING Blocked count: 0 Waited count: 381 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkUntil(LockSupport.java:410) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1726) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Thread 583 (ForkJoinPool.commonPool-worker-2): State: WAITING Blocked count: 0 Waited count: 43 Waiting on java.util.concurrent.ForkJoinPool@8b31de5 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1724) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Thread 604 (region-location-1): State: WAITING Blocked count: 6 Waited count: 11 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@71749958 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 605 (region-location-2): State: WAITING Blocked count: 4 Waited count: 9 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@71749958 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 606 (region-location-3): State: WAITING Blocked count: 3 Waited count: 8 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@71749958 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1007 (MutableQuantiles-0): State: TIMED_WAITING Blocked count: 0 Waited count: 420 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1071 (RPCClient-NioEventLoopGroup-6-2): State: RUNNABLE Blocked count: 26 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1111 (zk-permission-watcher-pool-0): State: WAITING Blocked count: 61 Waited count: 92 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@8cc91a Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1119 (RPCClient-NioEventLoopGroup-6-3): State: RUNNABLE Blocked count: 5 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1567 (Container metrics unregistration): State: WAITING Blocked count: 10 Waited count: 33 Waiting on java.util.TaskQueue@55e73e0a Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 1742 (ForkJoinPool.commonPool-worker-4): State: WAITING Blocked count: 0 Waited count: 584 Waiting on java.util.concurrent.ForkJoinPool@8b31de5 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1724) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Thread 2046 (RS-EventLoopGroup-3-2): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 2058 (RPCClient-NioEventLoopGroup-6-4): State: RUNNABLE Blocked count: 2 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 2059 (RS-EventLoopGroup-3-3): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 2420 (region-location-4): State: WAITING Blocked count: 2 Waited count: 7 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@71749958 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5232 (RPCClient-NioEventLoopGroup-6-5): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5233 (RPCClient-NioEventLoopGroup-6-6): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5234 (RPCClient-NioEventLoopGroup-6-7): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9459 (AsyncFSWAL-1-hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/MasterData-prefix:f28094b72178,42195,1736312287279): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@452b5d66 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9462 (Timer for 'JobHistoryServer' metrics system): State: TIMED_WAITING Blocked count: 0 Waited count: 12 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) 2025-01-08T05:06:05,795 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T05:06:35,795 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. Process Thread Dump: Automatic Stack Trace every 60 seconds waiting on M:0;f28094b72178:42195 220 active threads Thread 1 (main): State: TIMED_WAITING Blocked count: 3 Waited count: 4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.FutureTask.awaitDone(FutureTask.java:444) java.base@17.0.11/java.util.concurrent.FutureTask.get(FutureTask.java:203) app//org.junit.internal.runners.statements.FailOnTimeout.getResult(FailOnTimeout.java:167) app//org.junit.internal.runners.statements.FailOnTimeout.evaluate(FailOnTimeout.java:128) app//org.apache.hadoop.hbase.SystemExitRule$1.evaluate(SystemExitRule.java:39) app//org.junit.rules.RunRules.evaluate(RunRules.java:20) app//org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) app//org.junit.runners.ParentRunner.run(ParentRunner.java:413) app//org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:316) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:240) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:214) app//org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:155) app//org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:385) app//org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) app//org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:507) app//org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:495) Thread 2 (Reference Handler): State: RUNNABLE Blocked count: 9 Waited count: 0 Stack: java.base@17.0.11/java.lang.ref.Reference.waitForReferencePendingList(Native Method) java.base@17.0.11/java.lang.ref.Reference.processPendingReferences(Reference.java:253) java.base@17.0.11/java.lang.ref.Reference$ReferenceHandler.run(Reference.java:215) Thread 3 (Finalizer): State: WAITING Blocked count: 18 Waited count: 11 Waiting on java.lang.ref.ReferenceQueue$Lock@2d1af922 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) java.base@17.0.11/java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:172) Thread 4 (Signal Dispatcher): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 12 (Common-Cleaner): State: TIMED_WAITING Blocked count: 11 Waited count: 19 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/jdk.internal.ref.CleanerImpl.run(CleanerImpl.java:140) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) java.base@17.0.11/jdk.internal.misc.InnocuousThread.run(InnocuousThread.java:162) Thread 13 (Notification Thread): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 14 (pool-1-thread-1): State: RUNNABLE Blocked count: 0 Waited count: 23 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.poll(EPollPort.java:200) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:281) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 15 (pool-1-thread-2): State: WAITING Blocked count: 0 Waited count: 31 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5ff037ac Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:275) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 16 (surefire-forkedjvm-stream-flusher): State: TIMED_WAITING Blocked count: 0 Waited count: 5171 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 18 (surefire-forkedjvm-command-thread): State: WAITING Blocked count: 0 Waited count: 53 Waiting on java.util.concurrent.CountDownLatch$Sync@76ad1274 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.CountDownLatch.await(CountDownLatch.java:230) java.base@17.0.11/sun.nio.ch.PendingFuture.get(PendingFuture.java:178) app//org.apache.maven.surefire.api.util.internal.Channels$2.read(Channels.java:127) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:284) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) app//org.apache.maven.surefire.api.util.internal.Channels$3.readImpl(Channels.java:169) app//org.apache.maven.surefire.api.util.internal.AbstractNoninterruptibleReadableChannel.read(AbstractNoninterruptibleReadableChannel.java:50) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:430) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:419) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.readMessageType(AbstractStreamDecoder.java:116) app//org.apache.maven.surefire.booter.stream.CommandDecoder.decode(CommandDecoder.java:77) app//org.apache.maven.surefire.booter.spi.CommandChannelDecoder.decode(CommandChannelDecoder.java:60) app//org.apache.maven.surefire.booter.CommandReader$CommandRunnable.run(CommandReader.java:290) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 22 (Time-limited test): State: RUNNABLE Blocked count: 11521 Waited count: 12104 Stack: java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo1(Native Method) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:197) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:154) app//org.apache.hadoop.hbase.util.ReflectionUtils.printThreadInfo(ReflectionUtils.java:181) app//org.apache.hadoop.hbase.util.Threads.printThreadInfo(Threads.java:186) app//org.apache.hadoop.hbase.util.Threads.threadDumpingIsAlive(Threads.java:113) app//org.apache.hadoop.hbase.LocalHBaseCluster.join(LocalHBaseCluster.java:394) app//org.apache.hadoop.hbase.MiniHBaseCluster.waitUntilShutDown(MiniHBaseCluster.java:921) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniHBaseCluster(HBaseTestingUtility.java:1359) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniCluster(HBaseTestingUtility.java:1341) app//org.apache.hadoop.hbase.snapshot.TestExportSnapshot.tearDownAfterClass(TestExportSnapshot.java:121) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) java.base@17.0.11/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) java.base@17.0.11/java.lang.reflect.Method.invoke(Method.java:568) app//org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) app//org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) app//org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) app//org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) app//org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) Thread 23 (org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner): State: WAITING Blocked count: 9 Waited count: 10 Waiting on java.lang.ref.ReferenceQueue$Lock@3ae830cc Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) app//org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner.run(FileSystem.java:4171) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 25 (SSL Certificates Store Monitor): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.TaskQueue@647b2d6a Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 34 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@5eeedeb8): State: TIMED_WAITING Blocked count: 0 Waited count: 1028 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 35 (GcTimeMonitor obsWindow = 60000, sleepInterval = 5000, maxGcTimePerc = 100): State: TIMED_WAITING Blocked count: 0 Waited count: 103 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.GcTimeMonitor.run(GcTimeMonitor.java:161) Thread 36 (pool-6-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 37 (qtp338304781-37): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 38 (qtp338304781-38): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 39 (qtp338304781-39): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 40 (qtp338304781-40): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 41 (qtp338304781-41-acceptor-0@622697b3-ServerConnector@61d1e83b{HTTP/1.1, (http/1.1)}{localhost:40771}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 42 (qtp338304781-42): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 43 (qtp338304781-43): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 44 (qtp338304781-44): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 45 (Session-HouseKeeper-27700339-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 46 (pool-7-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 53 (FSEditLogAsync): State: WAITING Blocked count: 22 Waited count: 3310 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@75564eb2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.dequeueEdit(FSEditLogAsync.java:241) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.run(FSEditLogAsync.java:250) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 55 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 56 (IPC Server idle connection scanner for port 45939): State: TIMED_WAITING Blocked count: 1 Waited count: 53 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 58 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 103 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 61 (org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor@35667e22): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor.run(PendingReconstructionBlocks.java:267) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 62 (DatanodeAdminMonitor-0): State: TIMED_WAITING Blocked count: 0 Waited count: 171 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 49 (org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor@413afc06): State: TIMED_WAITING Blocked count: 0 Waited count: 103 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor.run(HeartbeatManager.java:563) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 47 (RedundancyMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 172 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) java.base@17.0.11/java.lang.Thread.sleep(Thread.java:344) java.base@17.0.11/java.util.concurrent.TimeUnit.sleep(TimeUnit.java:446) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$RedundancyMonitor.run(BlockManager.java:5352) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 48 (MarkedDeleteBlockScrubberThread): State: TIMED_WAITING Blocked count: 0 Waited count: 50176 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$MarkedDeleteBlockScrubber.run(BlockManager.java:5326) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 51 (Block report processor): State: WAITING Blocked count: 1 Waited count: 1555 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5e061a22 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.processQueue(BlockManager.java:5627) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.run(BlockManager.java:5614) Thread 57 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 54 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 64 (IPC Server handler 0 on default port 45939): State: TIMED_WAITING Blocked count: 74 Waited count: 2458 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 65 (IPC Server handler 1 on default port 45939): State: TIMED_WAITING Blocked count: 79 Waited count: 2479 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 66 (IPC Server handler 2 on default port 45939): State: TIMED_WAITING Blocked count: 70 Waited count: 2459 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 67 (IPC Server handler 3 on default port 45939): State: TIMED_WAITING Blocked count: 98 Waited count: 2473 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 68 (IPC Server handler 4 on default port 45939): State: TIMED_WAITING Blocked count: 74 Waited count: 2467 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 69 (pool-12-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 71 (org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor@32d56002): State: TIMED_WAITING Blocked count: 0 Waited count: 257 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor.run(LeaseManager.java:537) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 72 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor@351f5e8c): State: TIMED_WAITING Blocked count: 0 Waited count: 103 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor.run(FSNamesystem.java:4550) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 73 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller@40208336): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller.run(FSNamesystem.java:4592) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 74 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber@1d776204): State: TIMED_WAITING Blocked count: 0 Waited count: 3 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber.run(FSNamesystem.java:4689) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 75 (CacheReplicationMonitor(1494370843)): State: TIMED_WAITING Blocked count: 0 Waited count: 19 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1759) app//org.apache.hadoop.hdfs.server.blockmanagement.CacheReplicationMonitor.run(CacheReplicationMonitor.java:186) Thread 86 (pool-18-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 87 (qtp12592271-87): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 88 (qtp12592271-88-acceptor-0@1c732bcf-ServerConnector@7455429{HTTP/1.1, (http/1.1)}{localhost:33939}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 89 (qtp12592271-89): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 90 (qtp12592271-90): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 91 (Session-HouseKeeper-6dd22c37-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 92 (nioEventLoopGroup-2-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 93 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@39fef474): State: TIMED_WAITING Blocked count: 0 Waited count: 1025 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 95 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 96 (IPC Server idle connection scanner for port 45943): State: TIMED_WAITING Blocked count: 1 Waited count: 53 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 98 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 103 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 101 (Command processor): State: WAITING Blocked count: 2 Waited count: 300 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@29a0b510 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 102 (BP-1446750426-172.17.0.3-1736312281490 heartbeating to localhost/127.0.0.1:45939): State: TIMED_WAITING Blocked count: 1420 Waited count: 1629 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 103 (pool-20-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 85 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@4f29368a): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 97 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 94 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 104 (IPC Server handler 0 on default port 45943): State: TIMED_WAITING Blocked count: 0 Waited count: 514 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 105 (IPC Server handler 1 on default port 45943): State: TIMED_WAITING Blocked count: 0 Waited count: 513 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 106 (IPC Server handler 2 on default port 45943): State: TIMED_WAITING Blocked count: 0 Waited count: 513 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 107 (IPC Server handler 3 on default port 45943): State: TIMED_WAITING Blocked count: 0 Waited count: 513 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 108 (IPC Server handler 4 on default port 45943): State: TIMED_WAITING Blocked count: 0 Waited count: 513 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 120 (pool-26-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 121 (qtp1040803327-121): State: RUNNABLE Blocked count: 1 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 122 (qtp1040803327-122-acceptor-0@7588b8a-ServerConnector@673672ee{HTTP/1.1, (http/1.1)}{localhost:39721}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 118 (IPC Client (366384994) connection to localhost/127.0.0.1:45939 from jenkins): State: TIMED_WAITING Blocked count: 1485 Waited count: 1482 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Thread 123 (qtp1040803327-123): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 124 (qtp1040803327-124): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 119 (IPC Parameter Sending Thread for localhost/127.0.0.1:45939): State: TIMED_WAITING Blocked count: 0 Waited count: 2068 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 125 (Session-HouseKeeper-425819f9-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 126 (nioEventLoopGroup-4-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 127 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@49a955f5): State: TIMED_WAITING Blocked count: 0 Waited count: 1024 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 129 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 130 (IPC Server idle connection scanner for port 36851): State: TIMED_WAITING Blocked count: 1 Waited count: 53 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 132 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 103 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 137 (Command processor): State: WAITING Blocked count: 1 Waited count: 289 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@50b25eec Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 138 (BP-1446750426-172.17.0.3-1736312281490 heartbeating to localhost/127.0.0.1:45939): State: TIMED_WAITING Blocked count: 1442 Waited count: 1613 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 139 (pool-29-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 117 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@710dca53): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 131 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 128 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 140 (IPC Server handler 0 on default port 36851): State: TIMED_WAITING Blocked count: 0 Waited count: 512 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 141 (IPC Server handler 1 on default port 36851): State: TIMED_WAITING Blocked count: 0 Waited count: 512 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 142 (IPC Server handler 2 on default port 36851): State: TIMED_WAITING Blocked count: 0 Waited count: 512 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 143 (IPC Server handler 3 on default port 36851): State: TIMED_WAITING Blocked count: 0 Waited count: 512 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 144 (IPC Server handler 4 on default port 36851): State: TIMED_WAITING Blocked count: 0 Waited count: 512 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 154 (pool-36-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 156 (qtp244728962-156): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 157 (qtp244728962-157-acceptor-0@7ea21e77-ServerConnector@22e0a731{HTTP/1.1, (http/1.1)}{localhost:40719}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 158 (qtp244728962-158): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 159 (qtp244728962-159): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 160 (Session-HouseKeeper-562012e4-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 162 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data1)): State: TIMED_WAITING Blocked count: 2 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 164 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data2)): State: TIMED_WAITING Blocked count: 4 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 163 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data3)): State: TIMED_WAITING Blocked count: 25 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 165 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data4)): State: TIMED_WAITING Blocked count: 9 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 173 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data3/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 176 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data1/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 175 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data2/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 177 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data4/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 184 (nioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 188 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@1badcc77): State: TIMED_WAITING Blocked count: 0 Waited count: 1023 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 193 (pool-23-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 190 (pool-15-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 192 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 200 (IPC Server idle connection scanner for port 38935): State: TIMED_WAITING Blocked count: 1 Waited count: 53 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 199 (java.util.concurrent.ThreadPoolExecutor$Worker@3412cacb[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 196 (java.util.concurrent.ThreadPoolExecutor$Worker@17da2a7b[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 202 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 103 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 205 (Command processor): State: WAITING Blocked count: 2 Waited count: 330 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7c245a19 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 206 (BP-1446750426-172.17.0.3-1736312281490 heartbeating to localhost/127.0.0.1:45939): State: TIMED_WAITING Blocked count: 1394 Waited count: 1594 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 207 (pool-46-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 153 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@4ee4c44b): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 201 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 191 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 208 (IPC Server handler 0 on default port 38935): State: TIMED_WAITING Blocked count: 0 Waited count: 514 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 209 (IPC Server handler 1 on default port 38935): State: TIMED_WAITING Blocked count: 0 Waited count: 512 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 210 (IPC Server handler 2 on default port 38935): State: TIMED_WAITING Blocked count: 0 Waited count: 512 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 211 (IPC Server handler 3 on default port 38935): State: TIMED_WAITING Blocked count: 0 Waited count: 512 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 213 (IPC Server handler 4 on default port 38935): State: TIMED_WAITING Blocked count: 0 Waited count: 512 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 215 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data5)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 216 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data6)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 220 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data5/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 222 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data6/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 226 (pool-33-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 229 (java.util.concurrent.ThreadPoolExecutor$Worker@3f6c1323[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 234 (FsDatasetAsyncDiskServiceFixer): State: TIMED_WAITING Blocked count: 0 Waited count: 18 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer.run(HBaseTestingUtility.java:599) Thread 236 (NIOServerCxnFactory.SelectorThread-0): State: RUNNABLE Blocked count: 4 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 237 (NIOServerCxnFactory.SelectorThread-1): State: RUNNABLE Blocked count: 10 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 238 (NIOServerCxnFactory.AcceptThread:localhost/127.0.0.1:59048): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.select(NIOServerCnxnFactory.java:205) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.run(NIOServerCnxnFactory.java:181) Thread 235 (ConnnectionExpirer): State: TIMED_WAITING Blocked count: 0 Waited count: 52 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.NIOServerCnxnFactory$ConnectionExpirerThread.run(NIOServerCnxnFactory.java:554) Thread 239 (SessionTracker): State: TIMED_WAITING Blocked count: 0 Waited count: 256 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.SessionTrackerImpl.run(SessionTrackerImpl.java:163) Thread 240 (SyncThread:0): State: WAITING Blocked count: 22 Waited count: 763 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@51109921 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.SyncRequestProcessor.run(SyncRequestProcessor.java:170) Thread 241 (ProcessThread(sid:0 cport:59048):): State: WAITING Blocked count: 1 Waited count: 867 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5a02804a Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.PrepRequestProcessor.run(PrepRequestProcessor.java:142) Thread 242 (RequestThrottler): State: WAITING Blocked count: 0 Waited count: 891 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5cd4ed15 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.RequestThrottler.run(RequestThrottler.java:147) Thread 243 (NIOWorkerThread-1): State: WAITING Blocked count: 3 Waited count: 137 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 254 (Time-limited test.named-queue-events-pool-0): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@543bccbd Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) app//com.lmax.disruptor.BlockingWaitStrategy.waitFor(BlockingWaitStrategy.java:47) app//com.lmax.disruptor.ProcessingSequenceBarrier.waitFor(ProcessingSequenceBarrier.java:56) app//com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:159) app//com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 255 (HBase-Metrics2-1): State: TIMED_WAITING Blocked count: 0 Waited count: 389 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 256 (RS-EventLoopGroup-1-1): State: RUNNABLE Blocked count: 21 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 257 (Time-limited test-SendThread(127.0.0.1:59048)): State: RUNNABLE Blocked count: 42 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:332) app//org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1289) Thread 258 (Time-limited test-EventThread): State: WAITING Blocked count: 1 Waited count: 59 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@61b2fcae Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:550) Thread 259 (NIOWorkerThread-2): State: WAITING Blocked count: 2 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 260 (NIOWorkerThread-3): State: WAITING Blocked count: 2 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 261 (NIOWorkerThread-4): State: WAITING Blocked count: 1 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 262 (zk-event-processor-pool-0): State: WAITING Blocked count: 38 Waited count: 91 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5c993aff Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 263 (NIOWorkerThread-5): State: WAITING Blocked count: 3 Waited count: 137 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 264 (NIOWorkerThread-6): State: WAITING Blocked count: 2 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 265 (NIOWorkerThread-7): State: WAITING Blocked count: 3 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 266 (NIOWorkerThread-8): State: WAITING Blocked count: 1 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 267 (NIOWorkerThread-9): State: WAITING Blocked count: 3 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 268 (NIOWorkerThread-10): State: WAITING Blocked count: 2 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 269 (NIOWorkerThread-11): State: WAITING Blocked count: 1 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 270 (NIOWorkerThread-12): State: WAITING Blocked count: 0 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 271 (NIOWorkerThread-13): State: WAITING Blocked count: 2 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 272 (NIOWorkerThread-14): State: WAITING Blocked count: 2 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 273 (NIOWorkerThread-15): State: WAITING Blocked count: 4 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 274 (NIOWorkerThread-16): State: WAITING Blocked count: 2 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 275 (RpcServer.default.FPBQ.Fifo.handler=0,queue=0,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@564d5bf5 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 276 (RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195): State: WAITING Blocked count: 109 Waited count: 415 Waiting on java.util.concurrent.Semaphore$NonfairSync@643f078a Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 277 (RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195): State: WAITING Blocked count: 203 Waited count: 823 Waiting on java.util.concurrent.Semaphore$NonfairSync@129ed02f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 278 (RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42195): State: WAITING Blocked count: 38 Waited count: 9161 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7c54e41 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 279 (RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@3b6b3fbf Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 280 (RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@3b6b3fbf Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 281 (RpcServer.replication.FPBQ.Fifo.handler=0,queue=0,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@27f47c0f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 282 (RpcServer.replication.FPBQ.Fifo.handler=1,queue=0,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@438fcf76 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 283 (RpcServer.replication.FPBQ.Fifo.handler=2,queue=0,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@6e953600 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 284 (RpcServer.metaPriority.FPBQ.Fifo.handler=0,queue=0,port=42195): State: WAITING Blocked count: 0 Waited count: 7 Waiting on java.util.concurrent.Semaphore$NonfairSync@ce17aaa Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 288 (RS-EventLoopGroup-3-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 308 (RS-EventLoopGroup-4-1): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 328 (RS-EventLoopGroup-5-1): State: RUNNABLE Blocked count: 68 Waited count: 3 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 286 (M:0;f28094b72178:42195): State: TIMED_WAITING Blocked count: 8 Waited count: 3268 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1759) app//org.apache.hadoop.hbase.regionserver.wal.SyncFuture.get(SyncFuture.java:169) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.blockOnSync(AbstractFSWAL.java:879) app//org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.doSync(AsyncFSWAL.java:671) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.lambda$sync$1(AbstractFSWAL.java:600) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL$$Lambda$957/0x00007ffa3cf1b800.run(Unknown Source) app//org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:600) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:590) app//org.apache.hadoop.hbase.regionserver.HRegion.doSyncOfUnflushedWALChanges(HRegion.java:2869) app//org.apache.hadoop.hbase.regionserver.HRegion.internalPrepareFlushCache(HRegion.java:2811) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2670) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2644) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2635) app//org.apache.hadoop.hbase.regionserver.HRegion.doClose(HRegion.java:1810) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1631) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1586) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1569) Thread 349 (Monitor thread for TaskMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 51 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.monitoring.TaskMonitor$MonitorRunnable.run(TaskMonitor.java:325) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 351 (master/f28094b72178:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 353 (master/f28094b72178:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 355 (org.apache.hadoop.hdfs.PeerCache@7594f538): State: TIMED_WAITING Blocked count: 0 Waited count: 170 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.PeerCache.run(PeerCache.java:253) app//org.apache.hadoop.hdfs.PeerCache.access$000(PeerCache.java:46) app//org.apache.hadoop.hdfs.PeerCache$1.run(PeerCache.java:124) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 374 (master:store-WAL-Roller): State: TIMED_WAITING Blocked count: 0 Waited count: 5055 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hbase.wal.AbstractWALRoller.run(AbstractWALRoller.java:179) Thread 391 (RS-EventLoopGroup-5-2): State: RUNNABLE Blocked count: 53 Waited count: 4 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 392 (RS-EventLoopGroup-5-3): State: RUNNABLE Blocked count: 37 Waited count: 4 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 404 (Idle-Rpc-Conn-Sweeper-pool-0): State: WAITING Blocked count: 0 Waited count: 67 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@75e9f460 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 415 (SnapshotHandlerChoreCleaner): State: TIMED_WAITING Blocked count: 0 Waited count: 51 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 403 (RpcClient-timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 50497 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 424 (RS-EventLoopGroup-1-2): State: RUNNABLE Blocked count: 50 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 425 (RS-EventLoopGroup-1-3): State: RUNNABLE Blocked count: 34 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 440 (RegionServerTracker-0): State: WAITING Blocked count: 8 Waited count: 11 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4814fa0 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 471 (regionserver/f28094b72178:0.procedureResultReporter): State: WAITING Blocked count: 13 Waited count: 27 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@21db0ca7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 476 (regionserver/f28094b72178:0.procedureResultReporter): State: WAITING Blocked count: 19 Waited count: 39 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b2ba608 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 467 (regionserver/f28094b72178:0.procedureResultReporter): State: WAITING Blocked count: 9 Waited count: 19 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7f063557 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 512 (RS-EventLoopGroup-4-2): State: RUNNABLE Blocked count: 2 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 523 (RS-EventLoopGroup-4-3): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 525 (region-location-0): State: WAITING Blocked count: 7 Waited count: 12 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@71749958 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 551 (Async-Client-Retry-Timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 50231 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 565 (RPCClient-NioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 8 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 583 (ForkJoinPool.commonPool-worker-2): State: WAITING Blocked count: 0 Waited count: 43 Waiting on java.util.concurrent.ForkJoinPool@8b31de5 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1724) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Thread 604 (region-location-1): State: WAITING Blocked count: 6 Waited count: 11 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@71749958 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 605 (region-location-2): State: WAITING Blocked count: 4 Waited count: 9 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@71749958 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 606 (region-location-3): State: WAITING Blocked count: 3 Waited count: 8 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@71749958 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1007 (MutableQuantiles-0): State: TIMED_WAITING Blocked count: 0 Waited count: 426 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1071 (RPCClient-NioEventLoopGroup-6-2): State: RUNNABLE Blocked count: 26 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1111 (zk-permission-watcher-pool-0): State: WAITING Blocked count: 61 Waited count: 92 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@8cc91a Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1119 (RPCClient-NioEventLoopGroup-6-3): State: RUNNABLE Blocked count: 5 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1567 (Container metrics unregistration): State: WAITING Blocked count: 10 Waited count: 33 Waiting on java.util.TaskQueue@55e73e0a Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 1742 (ForkJoinPool.commonPool-worker-4): State: TIMED_WAITING Blocked count: 0 Waited count: 585 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkUntil(LockSupport.java:410) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1726) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Thread 2046 (RS-EventLoopGroup-3-2): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 2058 (RPCClient-NioEventLoopGroup-6-4): State: RUNNABLE Blocked count: 2 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 2059 (RS-EventLoopGroup-3-3): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 2420 (region-location-4): State: WAITING Blocked count: 2 Waited count: 7 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@71749958 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5232 (RPCClient-NioEventLoopGroup-6-5): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5233 (RPCClient-NioEventLoopGroup-6-6): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5234 (RPCClient-NioEventLoopGroup-6-7): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9459 (AsyncFSWAL-1-hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/MasterData-prefix:f28094b72178,42195,1736312287279): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@452b5d66 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9462 (Timer for 'JobHistoryServer' metrics system): State: TIMED_WAITING Blocked count: 0 Waited count: 18 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) 2025-01-08T05:07:05,796 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T05:07:35,796 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. Process Thread Dump: Automatic Stack Trace every 60 seconds waiting on M:0;f28094b72178:42195 220 active threads Thread 1 (main): State: TIMED_WAITING Blocked count: 3 Waited count: 4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.FutureTask.awaitDone(FutureTask.java:444) java.base@17.0.11/java.util.concurrent.FutureTask.get(FutureTask.java:203) app//org.junit.internal.runners.statements.FailOnTimeout.getResult(FailOnTimeout.java:167) app//org.junit.internal.runners.statements.FailOnTimeout.evaluate(FailOnTimeout.java:128) app//org.apache.hadoop.hbase.SystemExitRule$1.evaluate(SystemExitRule.java:39) app//org.junit.rules.RunRules.evaluate(RunRules.java:20) app//org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) app//org.junit.runners.ParentRunner.run(ParentRunner.java:413) app//org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:316) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:240) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:214) app//org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:155) app//org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:385) app//org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) app//org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:507) app//org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:495) Thread 2 (Reference Handler): State: RUNNABLE Blocked count: 9 Waited count: 0 Stack: java.base@17.0.11/java.lang.ref.Reference.waitForReferencePendingList(Native Method) java.base@17.0.11/java.lang.ref.Reference.processPendingReferences(Reference.java:253) java.base@17.0.11/java.lang.ref.Reference$ReferenceHandler.run(Reference.java:215) Thread 3 (Finalizer): State: WAITING Blocked count: 18 Waited count: 11 Waiting on java.lang.ref.ReferenceQueue$Lock@2d1af922 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) java.base@17.0.11/java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:172) Thread 4 (Signal Dispatcher): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 12 (Common-Cleaner): State: TIMED_WAITING Blocked count: 11 Waited count: 20 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/jdk.internal.ref.CleanerImpl.run(CleanerImpl.java:140) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) java.base@17.0.11/jdk.internal.misc.InnocuousThread.run(InnocuousThread.java:162) Thread 13 (Notification Thread): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 14 (pool-1-thread-1): State: RUNNABLE Blocked count: 0 Waited count: 26 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.poll(EPollPort.java:200) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:281) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 15 (pool-1-thread-2): State: WAITING Blocked count: 0 Waited count: 34 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5ff037ac Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:275) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 16 (surefire-forkedjvm-stream-flusher): State: TIMED_WAITING Blocked count: 0 Waited count: 5771 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 18 (surefire-forkedjvm-command-thread): State: WAITING Blocked count: 0 Waited count: 59 Waiting on java.util.concurrent.CountDownLatch$Sync@69f2b1aa Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.CountDownLatch.await(CountDownLatch.java:230) java.base@17.0.11/sun.nio.ch.PendingFuture.get(PendingFuture.java:178) app//org.apache.maven.surefire.api.util.internal.Channels$2.read(Channels.java:127) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:284) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) app//org.apache.maven.surefire.api.util.internal.Channels$3.readImpl(Channels.java:169) app//org.apache.maven.surefire.api.util.internal.AbstractNoninterruptibleReadableChannel.read(AbstractNoninterruptibleReadableChannel.java:50) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:430) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:419) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.readMessageType(AbstractStreamDecoder.java:116) app//org.apache.maven.surefire.booter.stream.CommandDecoder.decode(CommandDecoder.java:77) app//org.apache.maven.surefire.booter.spi.CommandChannelDecoder.decode(CommandChannelDecoder.java:60) app//org.apache.maven.surefire.booter.CommandReader$CommandRunnable.run(CommandReader.java:290) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 22 (Time-limited test): State: RUNNABLE Blocked count: 11521 Waited count: 12105 Stack: java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo1(Native Method) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:197) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:154) app//org.apache.hadoop.hbase.util.ReflectionUtils.printThreadInfo(ReflectionUtils.java:181) app//org.apache.hadoop.hbase.util.Threads.printThreadInfo(Threads.java:186) app//org.apache.hadoop.hbase.util.Threads.threadDumpingIsAlive(Threads.java:113) app//org.apache.hadoop.hbase.LocalHBaseCluster.join(LocalHBaseCluster.java:394) app//org.apache.hadoop.hbase.MiniHBaseCluster.waitUntilShutDown(MiniHBaseCluster.java:921) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniHBaseCluster(HBaseTestingUtility.java:1359) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniCluster(HBaseTestingUtility.java:1341) app//org.apache.hadoop.hbase.snapshot.TestExportSnapshot.tearDownAfterClass(TestExportSnapshot.java:121) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) java.base@17.0.11/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) java.base@17.0.11/java.lang.reflect.Method.invoke(Method.java:568) app//org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) app//org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) app//org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) app//org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) app//org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) Thread 23 (org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner): State: WAITING Blocked count: 9 Waited count: 10 Waiting on java.lang.ref.ReferenceQueue$Lock@3ae830cc Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) app//org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner.run(FileSystem.java:4171) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 25 (SSL Certificates Store Monitor): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.TaskQueue@647b2d6a Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 34 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@5eeedeb8): State: TIMED_WAITING Blocked count: 0 Waited count: 1148 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 35 (GcTimeMonitor obsWindow = 60000, sleepInterval = 5000, maxGcTimePerc = 100): State: TIMED_WAITING Blocked count: 0 Waited count: 115 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.GcTimeMonitor.run(GcTimeMonitor.java:161) Thread 36 (pool-6-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 37 (qtp338304781-37): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 38 (qtp338304781-38): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 39 (qtp338304781-39): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 40 (qtp338304781-40): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 41 (qtp338304781-41-acceptor-0@622697b3-ServerConnector@61d1e83b{HTTP/1.1, (http/1.1)}{localhost:40771}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 42 (qtp338304781-42): State: TIMED_WAITING Blocked count: 0 Waited count: 10 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 43 (qtp338304781-43): State: TIMED_WAITING Blocked count: 0 Waited count: 10 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 44 (qtp338304781-44): State: TIMED_WAITING Blocked count: 0 Waited count: 10 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 45 (Session-HouseKeeper-27700339-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 46 (pool-7-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 53 (FSEditLogAsync): State: WAITING Blocked count: 22 Waited count: 3310 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@75564eb2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.dequeueEdit(FSEditLogAsync.java:241) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.run(FSEditLogAsync.java:250) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 55 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 56 (IPC Server idle connection scanner for port 45939): State: TIMED_WAITING Blocked count: 1 Waited count: 59 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 58 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 115 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 61 (org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor@35667e22): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor.run(PendingReconstructionBlocks.java:267) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 62 (DatanodeAdminMonitor-0): State: TIMED_WAITING Blocked count: 0 Waited count: 191 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 49 (org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor@413afc06): State: TIMED_WAITING Blocked count: 0 Waited count: 115 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor.run(HeartbeatManager.java:563) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 47 (RedundancyMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 192 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) java.base@17.0.11/java.lang.Thread.sleep(Thread.java:344) java.base@17.0.11/java.util.concurrent.TimeUnit.sleep(TimeUnit.java:446) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$RedundancyMonitor.run(BlockManager.java:5352) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 48 (MarkedDeleteBlockScrubberThread): State: TIMED_WAITING Blocked count: 0 Waited count: 56136 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$MarkedDeleteBlockScrubber.run(BlockManager.java:5326) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 51 (Block report processor): State: WAITING Blocked count: 1 Waited count: 1555 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5e061a22 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.processQueue(BlockManager.java:5627) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.run(BlockManager.java:5614) Thread 57 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 54 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 64 (IPC Server handler 0 on default port 45939): State: TIMED_WAITING Blocked count: 80 Waited count: 2519 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 65 (IPC Server handler 1 on default port 45939): State: TIMED_WAITING Blocked count: 84 Waited count: 2540 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 66 (IPC Server handler 2 on default port 45939): State: TIMED_WAITING Blocked count: 74 Waited count: 2522 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 67 (IPC Server handler 3 on default port 45939): State: TIMED_WAITING Blocked count: 100 Waited count: 2534 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 68 (IPC Server handler 4 on default port 45939): State: TIMED_WAITING Blocked count: 77 Waited count: 2528 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 69 (pool-12-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 71 (org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor@32d56002): State: TIMED_WAITING Blocked count: 0 Waited count: 287 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor.run(LeaseManager.java:537) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 72 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor@351f5e8c): State: TIMED_WAITING Blocked count: 0 Waited count: 115 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor.run(FSNamesystem.java:4550) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 73 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller@40208336): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller.run(FSNamesystem.java:4592) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 74 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber@1d776204): State: TIMED_WAITING Blocked count: 0 Waited count: 3 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber.run(FSNamesystem.java:4689) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 75 (CacheReplicationMonitor(1494370843)): State: TIMED_WAITING Blocked count: 0 Waited count: 21 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1759) app//org.apache.hadoop.hdfs.server.blockmanagement.CacheReplicationMonitor.run(CacheReplicationMonitor.java:186) Thread 86 (pool-18-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 87 (qtp12592271-87): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 88 (qtp12592271-88-acceptor-0@1c732bcf-ServerConnector@7455429{HTTP/1.1, (http/1.1)}{localhost:33939}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 89 (qtp12592271-89): State: TIMED_WAITING Blocked count: 0 Waited count: 10 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 90 (qtp12592271-90): State: TIMED_WAITING Blocked count: 0 Waited count: 10 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 91 (Session-HouseKeeper-6dd22c37-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 92 (nioEventLoopGroup-2-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 93 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@39fef474): State: TIMED_WAITING Blocked count: 0 Waited count: 1145 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 95 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 96 (IPC Server idle connection scanner for port 45943): State: TIMED_WAITING Blocked count: 1 Waited count: 59 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 98 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 115 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 101 (Command processor): State: WAITING Blocked count: 2 Waited count: 320 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@29a0b510 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 102 (BP-1446750426-172.17.0.3-1736312281490 heartbeating to localhost/127.0.0.1:45939): State: TIMED_WAITING Blocked count: 1443 Waited count: 1681 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 103 (pool-20-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 85 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@4f29368a): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 97 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 94 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 104 (IPC Server handler 0 on default port 45943): State: TIMED_WAITING Blocked count: 0 Waited count: 574 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 105 (IPC Server handler 1 on default port 45943): State: TIMED_WAITING Blocked count: 0 Waited count: 573 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 106 (IPC Server handler 2 on default port 45943): State: TIMED_WAITING Blocked count: 0 Waited count: 573 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 107 (IPC Server handler 3 on default port 45943): State: TIMED_WAITING Blocked count: 0 Waited count: 573 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 108 (IPC Server handler 4 on default port 45943): State: TIMED_WAITING Blocked count: 0 Waited count: 573 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 120 (pool-26-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 121 (qtp1040803327-121): State: RUNNABLE Blocked count: 1 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 122 (qtp1040803327-122-acceptor-0@7588b8a-ServerConnector@673672ee{HTTP/1.1, (http/1.1)}{localhost:39721}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 118 (IPC Client (366384994) connection to localhost/127.0.0.1:45939 from jenkins): State: TIMED_WAITING Blocked count: 1525 Waited count: 1522 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Thread 123 (qtp1040803327-123): State: TIMED_WAITING Blocked count: 0 Waited count: 10 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 124 (qtp1040803327-124): State: TIMED_WAITING Blocked count: 0 Waited count: 10 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 119 (IPC Parameter Sending Thread for localhost/127.0.0.1:45939): State: TIMED_WAITING Blocked count: 0 Waited count: 2109 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 125 (Session-HouseKeeper-425819f9-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 126 (nioEventLoopGroup-4-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 127 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@49a955f5): State: TIMED_WAITING Blocked count: 0 Waited count: 1144 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 129 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 130 (IPC Server idle connection scanner for port 36851): State: TIMED_WAITING Blocked count: 1 Waited count: 59 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 132 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 115 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 137 (Command processor): State: WAITING Blocked count: 1 Waited count: 309 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@50b25eec Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 138 (BP-1446750426-172.17.0.3-1736312281490 heartbeating to localhost/127.0.0.1:45939): State: TIMED_WAITING Blocked count: 1463 Waited count: 1660 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 139 (pool-29-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 117 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@710dca53): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 131 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 128 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 140 (IPC Server handler 0 on default port 36851): State: TIMED_WAITING Blocked count: 0 Waited count: 572 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 141 (IPC Server handler 1 on default port 36851): State: TIMED_WAITING Blocked count: 0 Waited count: 572 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 142 (IPC Server handler 2 on default port 36851): State: TIMED_WAITING Blocked count: 0 Waited count: 572 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 143 (IPC Server handler 3 on default port 36851): State: TIMED_WAITING Blocked count: 0 Waited count: 572 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 144 (IPC Server handler 4 on default port 36851): State: TIMED_WAITING Blocked count: 0 Waited count: 572 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 154 (pool-36-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 156 (qtp244728962-156): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 157 (qtp244728962-157-acceptor-0@7ea21e77-ServerConnector@22e0a731{HTTP/1.1, (http/1.1)}{localhost:40719}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 158 (qtp244728962-158): State: TIMED_WAITING Blocked count: 0 Waited count: 10 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 159 (qtp244728962-159): State: TIMED_WAITING Blocked count: 0 Waited count: 10 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 160 (Session-HouseKeeper-562012e4-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 162 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data1)): State: TIMED_WAITING Blocked count: 2 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 164 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data2)): State: TIMED_WAITING Blocked count: 4 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 163 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data3)): State: TIMED_WAITING Blocked count: 25 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 165 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data4)): State: TIMED_WAITING Blocked count: 9 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 173 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data3/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 176 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data1/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 2 Waited count: 3 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 175 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data2/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 177 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data4/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 184 (nioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 188 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@1badcc77): State: TIMED_WAITING Blocked count: 0 Waited count: 1143 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 193 (pool-23-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 190 (pool-15-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 192 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 200 (IPC Server idle connection scanner for port 38935): State: TIMED_WAITING Blocked count: 1 Waited count: 59 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 199 (java.util.concurrent.ThreadPoolExecutor$Worker@3412cacb[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 196 (java.util.concurrent.ThreadPoolExecutor$Worker@17da2a7b[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 202 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 115 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 205 (Command processor): State: WAITING Blocked count: 2 Waited count: 350 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7c245a19 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 206 (BP-1446750426-172.17.0.3-1736312281490 heartbeating to localhost/127.0.0.1:45939): State: TIMED_WAITING Blocked count: 1414 Waited count: 1634 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 207 (pool-46-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 153 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@4ee4c44b): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 201 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 191 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 208 (IPC Server handler 0 on default port 38935): State: TIMED_WAITING Blocked count: 0 Waited count: 574 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 209 (IPC Server handler 1 on default port 38935): State: TIMED_WAITING Blocked count: 0 Waited count: 572 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 210 (IPC Server handler 2 on default port 38935): State: TIMED_WAITING Blocked count: 0 Waited count: 572 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 211 (IPC Server handler 3 on default port 38935): State: TIMED_WAITING Blocked count: 0 Waited count: 572 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 213 (IPC Server handler 4 on default port 38935): State: TIMED_WAITING Blocked count: 0 Waited count: 572 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 215 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data5)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 216 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data6)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 220 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data5/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 2 Waited count: 3 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 222 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data6/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 226 (pool-33-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 229 (java.util.concurrent.ThreadPoolExecutor$Worker@3f6c1323[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 234 (FsDatasetAsyncDiskServiceFixer): State: TIMED_WAITING Blocked count: 0 Waited count: 20 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer.run(HBaseTestingUtility.java:599) Thread 236 (NIOServerCxnFactory.SelectorThread-0): State: RUNNABLE Blocked count: 4 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 237 (NIOServerCxnFactory.SelectorThread-1): State: RUNNABLE Blocked count: 10 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 238 (NIOServerCxnFactory.AcceptThread:localhost/127.0.0.1:59048): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.select(NIOServerCnxnFactory.java:205) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.run(NIOServerCnxnFactory.java:181) Thread 235 (ConnnectionExpirer): State: TIMED_WAITING Blocked count: 0 Waited count: 58 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.NIOServerCnxnFactory$ConnectionExpirerThread.run(NIOServerCnxnFactory.java:554) Thread 239 (SessionTracker): State: TIMED_WAITING Blocked count: 0 Waited count: 286 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.SessionTrackerImpl.run(SessionTrackerImpl.java:163) Thread 240 (SyncThread:0): State: WAITING Blocked count: 22 Waited count: 767 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@51109921 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.SyncRequestProcessor.run(SyncRequestProcessor.java:170) Thread 241 (ProcessThread(sid:0 cport:59048):): State: WAITING Blocked count: 1 Waited count: 871 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5a02804a Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.PrepRequestProcessor.run(PrepRequestProcessor.java:142) Thread 242 (RequestThrottler): State: WAITING Blocked count: 0 Waited count: 895 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5cd4ed15 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.RequestThrottler.run(RequestThrottler.java:147) Thread 243 (NIOWorkerThread-1): State: WAITING Blocked count: 3 Waited count: 137 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 254 (Time-limited test.named-queue-events-pool-0): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@543bccbd Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) app//com.lmax.disruptor.BlockingWaitStrategy.waitFor(BlockingWaitStrategy.java:47) app//com.lmax.disruptor.ProcessingSequenceBarrier.waitFor(ProcessingSequenceBarrier.java:56) app//com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:159) app//com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 255 (HBase-Metrics2-1): State: TIMED_WAITING Blocked count: 0 Waited count: 417 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 256 (RS-EventLoopGroup-1-1): State: RUNNABLE Blocked count: 21 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 257 (Time-limited test-SendThread(127.0.0.1:59048)): State: RUNNABLE Blocked count: 42 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:332) app//org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1289) Thread 258 (Time-limited test-EventThread): State: WAITING Blocked count: 1 Waited count: 59 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@61b2fcae Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:550) Thread 259 (NIOWorkerThread-2): State: WAITING Blocked count: 2 Waited count: 137 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 260 (NIOWorkerThread-3): State: WAITING Blocked count: 2 Waited count: 137 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 261 (NIOWorkerThread-4): State: WAITING Blocked count: 1 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 262 (zk-event-processor-pool-0): State: WAITING Blocked count: 38 Waited count: 91 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5c993aff Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 263 (NIOWorkerThread-5): State: WAITING Blocked count: 3 Waited count: 137 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 264 (NIOWorkerThread-6): State: WAITING Blocked count: 2 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 265 (NIOWorkerThread-7): State: WAITING Blocked count: 3 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 266 (NIOWorkerThread-8): State: WAITING Blocked count: 1 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 267 (NIOWorkerThread-9): State: WAITING Blocked count: 3 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 268 (NIOWorkerThread-10): State: WAITING Blocked count: 2 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 269 (NIOWorkerThread-11): State: WAITING Blocked count: 1 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 270 (NIOWorkerThread-12): State: WAITING Blocked count: 0 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 271 (NIOWorkerThread-13): State: WAITING Blocked count: 2 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 272 (NIOWorkerThread-14): State: WAITING Blocked count: 2 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 273 (NIOWorkerThread-15): State: WAITING Blocked count: 4 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 274 (NIOWorkerThread-16): State: WAITING Blocked count: 2 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 275 (RpcServer.default.FPBQ.Fifo.handler=0,queue=0,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@564d5bf5 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 276 (RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195): State: WAITING Blocked count: 109 Waited count: 415 Waiting on java.util.concurrent.Semaphore$NonfairSync@643f078a Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 277 (RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195): State: WAITING Blocked count: 203 Waited count: 823 Waiting on java.util.concurrent.Semaphore$NonfairSync@129ed02f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 278 (RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42195): State: WAITING Blocked count: 38 Waited count: 9161 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7c54e41 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 279 (RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@3b6b3fbf Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 280 (RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@3b6b3fbf Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 281 (RpcServer.replication.FPBQ.Fifo.handler=0,queue=0,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@27f47c0f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 282 (RpcServer.replication.FPBQ.Fifo.handler=1,queue=0,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@438fcf76 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 283 (RpcServer.replication.FPBQ.Fifo.handler=2,queue=0,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@6e953600 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 284 (RpcServer.metaPriority.FPBQ.Fifo.handler=0,queue=0,port=42195): State: WAITING Blocked count: 0 Waited count: 7 Waiting on java.util.concurrent.Semaphore$NonfairSync@ce17aaa Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 288 (RS-EventLoopGroup-3-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 308 (RS-EventLoopGroup-4-1): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 328 (RS-EventLoopGroup-5-1): State: RUNNABLE Blocked count: 68 Waited count: 3 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 286 (M:0;f28094b72178:42195): State: TIMED_WAITING Blocked count: 8 Waited count: 3268 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1759) app//org.apache.hadoop.hbase.regionserver.wal.SyncFuture.get(SyncFuture.java:169) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.blockOnSync(AbstractFSWAL.java:879) app//org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.doSync(AsyncFSWAL.java:671) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.lambda$sync$1(AbstractFSWAL.java:600) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL$$Lambda$957/0x00007ffa3cf1b800.run(Unknown Source) app//org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:600) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:590) app//org.apache.hadoop.hbase.regionserver.HRegion.doSyncOfUnflushedWALChanges(HRegion.java:2869) app//org.apache.hadoop.hbase.regionserver.HRegion.internalPrepareFlushCache(HRegion.java:2811) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2670) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2644) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2635) app//org.apache.hadoop.hbase.regionserver.HRegion.doClose(HRegion.java:1810) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1631) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1586) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1569) Thread 349 (Monitor thread for TaskMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 57 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.monitoring.TaskMonitor$MonitorRunnable.run(TaskMonitor.java:325) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 351 (master/f28094b72178:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 353 (master/f28094b72178:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 355 (org.apache.hadoop.hdfs.PeerCache@7594f538): State: TIMED_WAITING Blocked count: 0 Waited count: 190 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.PeerCache.run(PeerCache.java:253) app//org.apache.hadoop.hdfs.PeerCache.access$000(PeerCache.java:46) app//org.apache.hadoop.hdfs.PeerCache$1.run(PeerCache.java:124) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 374 (master:store-WAL-Roller): State: TIMED_WAITING Blocked count: 0 Waited count: 5654 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hbase.wal.AbstractWALRoller.run(AbstractWALRoller.java:179) Thread 391 (RS-EventLoopGroup-5-2): State: RUNNABLE Blocked count: 53 Waited count: 4 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 392 (RS-EventLoopGroup-5-3): State: RUNNABLE Blocked count: 37 Waited count: 4 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 404 (Idle-Rpc-Conn-Sweeper-pool-0): State: WAITING Blocked count: 0 Waited count: 67 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@75e9f460 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 415 (SnapshotHandlerChoreCleaner): State: TIMED_WAITING Blocked count: 0 Waited count: 57 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 403 (RpcClient-timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 56498 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 424 (RS-EventLoopGroup-1-2): State: RUNNABLE Blocked count: 50 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 425 (RS-EventLoopGroup-1-3): State: RUNNABLE Blocked count: 34 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 440 (RegionServerTracker-0): State: WAITING Blocked count: 8 Waited count: 11 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4814fa0 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 471 (regionserver/f28094b72178:0.procedureResultReporter): State: WAITING Blocked count: 13 Waited count: 27 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@21db0ca7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 476 (regionserver/f28094b72178:0.procedureResultReporter): State: WAITING Blocked count: 19 Waited count: 39 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b2ba608 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 467 (regionserver/f28094b72178:0.procedureResultReporter): State: WAITING Blocked count: 9 Waited count: 19 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7f063557 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 512 (RS-EventLoopGroup-4-2): State: RUNNABLE Blocked count: 2 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 523 (RS-EventLoopGroup-4-3): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 525 (region-location-0): State: WAITING Blocked count: 7 Waited count: 12 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@71749958 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 551 (Async-Client-Retry-Timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 56233 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 565 (RPCClient-NioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 8 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 583 (ForkJoinPool.commonPool-worker-2): State: TIMED_WAITING Blocked count: 0 Waited count: 44 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkUntil(LockSupport.java:410) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1726) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Thread 604 (region-location-1): State: WAITING Blocked count: 6 Waited count: 11 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@71749958 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 605 (region-location-2): State: WAITING Blocked count: 4 Waited count: 9 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@71749958 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 606 (region-location-3): State: WAITING Blocked count: 3 Waited count: 8 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@71749958 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1007 (MutableQuantiles-0): State: TIMED_WAITING Blocked count: 0 Waited count: 432 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1071 (RPCClient-NioEventLoopGroup-6-2): State: RUNNABLE Blocked count: 26 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1111 (zk-permission-watcher-pool-0): State: WAITING Blocked count: 61 Waited count: 92 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@8cc91a Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1119 (RPCClient-NioEventLoopGroup-6-3): State: RUNNABLE Blocked count: 5 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1567 (Container metrics unregistration): State: WAITING Blocked count: 10 Waited count: 33 Waiting on java.util.TaskQueue@55e73e0a Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 2046 (RS-EventLoopGroup-3-2): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 2058 (RPCClient-NioEventLoopGroup-6-4): State: RUNNABLE Blocked count: 2 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 2059 (RS-EventLoopGroup-3-3): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 2420 (region-location-4): State: WAITING Blocked count: 2 Waited count: 7 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@71749958 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5232 (RPCClient-NioEventLoopGroup-6-5): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5233 (RPCClient-NioEventLoopGroup-6-6): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5234 (RPCClient-NioEventLoopGroup-6-7): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9459 (AsyncFSWAL-1-hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/MasterData-prefix:f28094b72178,42195,1736312287279): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@452b5d66 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9462 (Timer for 'JobHistoryServer' metrics system): State: TIMED_WAITING Blocked count: 0 Waited count: 24 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 9463 (process reaper): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:401) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) 2025-01-08T05:08:05,796 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T05:08:09,033 DEBUG [master/f28094b72178:0:becomeActiveMaster-MemStoreChunkPool Statistics {}] regionserver.ChunkCreator$MemStoreChunkPool$StatisticsThread(417): data stats (chunk size=2097152): current pool size=7, created chunk count=10, reused chunk count=22, reuseRatio=68.75% 2025-01-08T05:08:09,033 DEBUG [master/f28094b72178:0:becomeActiveMaster-MemStoreChunkPool Statistics {}] regionserver.ChunkCreator$MemStoreChunkPool$StatisticsThread(417): index stats (chunk size=209715): current pool size=0, created chunk count=0, reused chunk count=0, reuseRatio=0 2025-01-08T05:08:17,579 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T05:08:35,796 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T05:08:36,263 DEBUG [M:0;f28094b72178:42195 {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2025-01-08T05:08:36,263 WARN [M:0;f28094b72178:42195 {}] region.MasterRegion(134): Failed to close region org.apache.hadoop.hbase.regionserver.wal.WALSyncTimeoutIOException: org.apache.hadoop.hbase.exceptions.TimeoutIOException: Failed to get sync result after 300000 ms for txid=3778, WAL system stuck? at org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.blockOnSync(AbstractFSWAL.java:883) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.doSync(AsyncFSWAL.java:671) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.lambda$sync$1(AbstractFSWAL.java:600) ~[classes/:?] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) ~[hbase-common-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:600) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:590) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.doSyncOfUnflushedWALChanges(HRegion.java:2869) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.internalPrepareFlushCache(HRegion.java:2811) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2670) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2644) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2635) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.doClose(HRegion.java:1810) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1631) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1586) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1569) ~[classes/:?] at org.apache.hadoop.hbase.master.region.MasterRegion.closeRegion(MasterRegion.java:132) ~[classes/:?] at org.apache.hadoop.hbase.master.region.MasterRegion.close(MasterRegion.java:205) ~[classes/:?] at org.apache.hadoop.hbase.master.HMaster.stopServiceThreads(HMaster.java:1769) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:1285) ~[classes/:?] at org.apache.hadoop.hbase.master.HMaster.run(HMaster.java:614) ~[classes/:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: org.apache.hadoop.hbase.exceptions.TimeoutIOException: Failed to get sync result after 300000 ms for txid=3778, WAL system stuck? at org.apache.hadoop.hbase.regionserver.wal.SyncFuture.get(SyncFuture.java:171) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.blockOnSync(AbstractFSWAL.java:879) ~[classes/:?] ... 20 more 2025-01-08T05:08:36,264 WARN [Close-WAL-Writer-0 {}] wal.AsyncProtobufLogWriter(163): normal close failed, try recover java.lang.NullPointerException: Cannot invoke "org.apache.hbase.thirdparty.io.netty.buffer.ByteBuf.ensureWritable(int)" because "this.buf" is null at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutput.writeInt(FanOutOneBlockAsyncDFSOutput.java:396) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.wal.AsyncProtobufLogWriter.lambda$writeWALTrailerAndMagic$3(AsyncProtobufLogWriter.java:243) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncProtobufLogWriter.writeWALMetadata(AsyncProtobufLogWriter.java:201) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncProtobufLogWriter.writeWALTrailerAndMagic(AsyncProtobufLogWriter.java:236) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AbstractProtobufLogWriter.writeWALTrailer(AbstractProtobufLogWriter.java:252) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncProtobufLogWriter.close(AsyncProtobufLogWriter.java:160) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.lambda$closeWriter$5(AsyncFSWAL.java:732) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2025-01-08T05:08:36,266 DEBUG [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(47): Initialize RecoverLeaseFSUtils 2025-01-08T05:08:36,266 DEBUG [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(59): set recoverLeaseMethod to org.apache.hadoop.fs.LeaseRecoverable.recoverLease() 2025-01-08T05:08:36,266 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file /user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/MasterData/WALs/f28094b72178,42195,1736312287279/f28094b72178%2C42195%2C1736312287279.1736312289851 2025-01-08T05:08:36,267 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(233): attempt=0 on file=/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/MasterData/WALs/f28094b72178,42195,1736312287279/f28094b72178%2C42195%2C1736312287279.1736312289851 after 1ms java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.recoverLease(DFSClient.java:946) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:317) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:314) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.recoverLease(DistributedFileSystem.java:329) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverLease(RecoverLeaseFSUtils.java:222) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:155) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutput.recoverAndClose(FanOutOneBlockAsyncDFSOutput.java:610) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.wal.AsyncProtobufLogWriter.close(AsyncProtobufLogWriter.java:164) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.lambda$closeWriter$5(AsyncFSWAL.java:732) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2025-01-08T05:08:36,267 WARN [Close-WAL-Writer-0 {}] wal.AsyncFSWAL(734): close old writer failed. java.io.InterruptedIOException: Operation cancelled at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.checkIfCancelled(RecoverLeaseFSUtils.java:269) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:159) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutput.recoverAndClose(FanOutOneBlockAsyncDFSOutput.java:610) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.wal.AsyncProtobufLogWriter.close(AsyncProtobufLogWriter.java:164) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.lambda$closeWriter$5(AsyncFSWAL.java:732) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2025-01-08T05:08:36,267 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/MasterData/WALs/f28094b72178,42195,1736312287279/f28094b72178%2C42195%2C1736312287279.1736312289851 2025-01-08T05:08:36,267 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(233): attempt=0 on file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/MasterData/WALs/f28094b72178,42195,1736312287279/f28094b72178%2C42195%2C1736312287279.1736312289851 after 0ms java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.recoverLease(DFSClient.java:946) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:317) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:314) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.recoverLease(DistributedFileSystem.java:329) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverLease(RecoverLeaseFSUtils.java:222) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:155) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.recoverLease(AsyncFSWAL.java:722) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.lambda$closeWriter$5(AsyncFSWAL.java:735) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Process Thread Dump: Automatic Stack Trace every 60 seconds waiting on M:0;f28094b72178:42195 221 active threads Thread 1 (main): State: TIMED_WAITING Blocked count: 3 Waited count: 4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.FutureTask.awaitDone(FutureTask.java:444) java.base@17.0.11/java.util.concurrent.FutureTask.get(FutureTask.java:203) app//org.junit.internal.runners.statements.FailOnTimeout.getResult(FailOnTimeout.java:167) app//org.junit.internal.runners.statements.FailOnTimeout.evaluate(FailOnTimeout.java:128) app//org.apache.hadoop.hbase.SystemExitRule$1.evaluate(SystemExitRule.java:39) app//org.junit.rules.RunRules.evaluate(RunRules.java:20) app//org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) app//org.junit.runners.ParentRunner.run(ParentRunner.java:413) app//org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:316) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:240) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:214) app//org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:155) app//org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:385) app//org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) app//org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:507) app//org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:495) Thread 2 (Reference Handler): State: RUNNABLE Blocked count: 9 Waited count: 0 Stack: java.base@17.0.11/java.lang.ref.Reference.waitForReferencePendingList(Native Method) java.base@17.0.11/java.lang.ref.Reference.processPendingReferences(Reference.java:253) java.base@17.0.11/java.lang.ref.Reference$ReferenceHandler.run(Reference.java:215) Thread 3 (Finalizer): State: WAITING Blocked count: 18 Waited count: 11 Waiting on java.lang.ref.ReferenceQueue$Lock@2d1af922 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) java.base@17.0.11/java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:172) Thread 4 (Signal Dispatcher): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 12 (Common-Cleaner): State: TIMED_WAITING Blocked count: 11 Waited count: 21 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/jdk.internal.ref.CleanerImpl.run(CleanerImpl.java:140) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) java.base@17.0.11/jdk.internal.misc.InnocuousThread.run(InnocuousThread.java:162) Thread 13 (Notification Thread): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 14 (pool-1-thread-1): State: RUNNABLE Blocked count: 0 Waited count: 29 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.poll(EPollPort.java:200) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:281) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 15 (pool-1-thread-2): State: WAITING Blocked count: 0 Waited count: 37 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5ff037ac Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:275) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 16 (surefire-forkedjvm-stream-flusher): State: TIMED_WAITING Blocked count: 0 Waited count: 6370 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 18 (surefire-forkedjvm-command-thread): State: WAITING Blocked count: 0 Waited count: 65 Waiting on java.util.concurrent.CountDownLatch$Sync@1934a622 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.CountDownLatch.await(CountDownLatch.java:230) java.base@17.0.11/sun.nio.ch.PendingFuture.get(PendingFuture.java:178) app//org.apache.maven.surefire.api.util.internal.Channels$2.read(Channels.java:127) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:284) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) app//org.apache.maven.surefire.api.util.internal.Channels$3.readImpl(Channels.java:169) app//org.apache.maven.surefire.api.util.internal.AbstractNoninterruptibleReadableChannel.read(AbstractNoninterruptibleReadableChannel.java:50) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:430) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:419) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.readMessageType(AbstractStreamDecoder.java:116) app//org.apache.maven.surefire.booter.stream.CommandDecoder.decode(CommandDecoder.java:77) app//org.apache.maven.surefire.booter.spi.CommandChannelDecoder.decode(CommandChannelDecoder.java:60) app//org.apache.maven.surefire.booter.CommandReader$CommandRunnable.run(CommandReader.java:290) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 22 (Time-limited test): State: RUNNABLE Blocked count: 11521 Waited count: 12106 Stack: java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo1(Native Method) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:197) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:154) app//org.apache.hadoop.hbase.util.ReflectionUtils.printThreadInfo(ReflectionUtils.java:181) app//org.apache.hadoop.hbase.util.Threads.printThreadInfo(Threads.java:186) app//org.apache.hadoop.hbase.util.Threads.threadDumpingIsAlive(Threads.java:113) app//org.apache.hadoop.hbase.LocalHBaseCluster.join(LocalHBaseCluster.java:394) app//org.apache.hadoop.hbase.MiniHBaseCluster.waitUntilShutDown(MiniHBaseCluster.java:921) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniHBaseCluster(HBaseTestingUtility.java:1359) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniCluster(HBaseTestingUtility.java:1341) app//org.apache.hadoop.hbase.snapshot.TestExportSnapshot.tearDownAfterClass(TestExportSnapshot.java:121) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) java.base@17.0.11/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) java.base@17.0.11/java.lang.reflect.Method.invoke(Method.java:568) app//org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) app//org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) app//org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) app//org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) app//org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) Thread 23 (org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner): State: WAITING Blocked count: 9 Waited count: 10 Waiting on java.lang.ref.ReferenceQueue$Lock@3ae830cc Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) app//org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner.run(FileSystem.java:4171) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 25 (SSL Certificates Store Monitor): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.TaskQueue@647b2d6a Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 34 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@5eeedeb8): State: TIMED_WAITING Blocked count: 0 Waited count: 1268 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 35 (GcTimeMonitor obsWindow = 60000, sleepInterval = 5000, maxGcTimePerc = 100): State: TIMED_WAITING Blocked count: 0 Waited count: 127 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.GcTimeMonitor.run(GcTimeMonitor.java:161) Thread 36 (pool-6-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 37 (qtp338304781-37): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 38 (qtp338304781-38): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 39 (qtp338304781-39): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 40 (qtp338304781-40): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 41 (qtp338304781-41-acceptor-0@622697b3-ServerConnector@61d1e83b{HTTP/1.1, (http/1.1)}{localhost:40771}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 42 (qtp338304781-42): State: TIMED_WAITING Blocked count: 0 Waited count: 11 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 43 (qtp338304781-43): State: TIMED_WAITING Blocked count: 0 Waited count: 11 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 44 (qtp338304781-44): State: TIMED_WAITING Blocked count: 0 Waited count: 11 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 45 (Session-HouseKeeper-27700339-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 46 (pool-7-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 53 (FSEditLogAsync): State: WAITING Blocked count: 22 Waited count: 3310 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@75564eb2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.dequeueEdit(FSEditLogAsync.java:241) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.run(FSEditLogAsync.java:250) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 55 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 56 (IPC Server idle connection scanner for port 45939): State: TIMED_WAITING Blocked count: 1 Waited count: 65 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 58 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 127 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 61 (org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor@35667e22): State: TIMED_WAITING Blocked count: 0 Waited count: 3 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor.run(PendingReconstructionBlocks.java:267) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 62 (DatanodeAdminMonitor-0): State: TIMED_WAITING Blocked count: 0 Waited count: 211 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 49 (org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor@413afc06): State: TIMED_WAITING Blocked count: 0 Waited count: 127 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor.run(HeartbeatManager.java:563) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 47 (RedundancyMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 212 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) java.base@17.0.11/java.lang.Thread.sleep(Thread.java:344) java.base@17.0.11/java.util.concurrent.TimeUnit.sleep(TimeUnit.java:446) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$RedundancyMonitor.run(BlockManager.java:5352) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 48 (MarkedDeleteBlockScrubberThread): State: TIMED_WAITING Blocked count: 0 Waited count: 62064 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$MarkedDeleteBlockScrubber.run(BlockManager.java:5326) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 51 (Block report processor): State: WAITING Blocked count: 1 Waited count: 1555 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5e061a22 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.processQueue(BlockManager.java:5627) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.run(BlockManager.java:5614) Thread 57 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 54 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 64 (IPC Server handler 0 on default port 45939): State: TIMED_WAITING Blocked count: 82 Waited count: 2580 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 65 (IPC Server handler 1 on default port 45939): State: TIMED_WAITING Blocked count: 86 Waited count: 2601 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 66 (IPC Server handler 2 on default port 45939): State: TIMED_WAITING Blocked count: 74 Waited count: 2584 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 67 (IPC Server handler 3 on default port 45939): State: TIMED_WAITING Blocked count: 101 Waited count: 2598 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 68 (IPC Server handler 4 on default port 45939): State: TIMED_WAITING Blocked count: 77 Waited count: 2589 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 69 (pool-12-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 71 (org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor@32d56002): State: TIMED_WAITING Blocked count: 0 Waited count: 317 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor.run(LeaseManager.java:537) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 72 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor@351f5e8c): State: TIMED_WAITING Blocked count: 0 Waited count: 127 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor.run(FSNamesystem.java:4550) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 73 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller@40208336): State: TIMED_WAITING Blocked count: 0 Waited count: 3 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller.run(FSNamesystem.java:4592) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 74 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber@1d776204): State: TIMED_WAITING Blocked count: 0 Waited count: 4 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber.run(FSNamesystem.java:4689) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 75 (CacheReplicationMonitor(1494370843)): State: TIMED_WAITING Blocked count: 0 Waited count: 23 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1759) app//org.apache.hadoop.hdfs.server.blockmanagement.CacheReplicationMonitor.run(CacheReplicationMonitor.java:186) Thread 86 (pool-18-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 87 (qtp12592271-87): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 88 (qtp12592271-88-acceptor-0@1c732bcf-ServerConnector@7455429{HTTP/1.1, (http/1.1)}{localhost:33939}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 89 (qtp12592271-89): State: TIMED_WAITING Blocked count: 0 Waited count: 11 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 90 (qtp12592271-90): State: TIMED_WAITING Blocked count: 0 Waited count: 11 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 91 (Session-HouseKeeper-6dd22c37-1): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 92 (nioEventLoopGroup-2-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 93 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@39fef474): State: TIMED_WAITING Blocked count: 0 Waited count: 1264 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 95 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 96 (IPC Server idle connection scanner for port 45943): State: TIMED_WAITING Blocked count: 1 Waited count: 65 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 98 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 127 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 101 (Command processor): State: WAITING Blocked count: 2 Waited count: 340 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@29a0b510 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 102 (BP-1446750426-172.17.0.3-1736312281490 heartbeating to localhost/127.0.0.1:45939): State: TIMED_WAITING Blocked count: 1463 Waited count: 1724 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 103 (pool-20-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 85 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@4f29368a): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 97 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 94 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 104 (IPC Server handler 0 on default port 45943): State: TIMED_WAITING Blocked count: 0 Waited count: 634 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 105 (IPC Server handler 1 on default port 45943): State: TIMED_WAITING Blocked count: 0 Waited count: 633 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 106 (IPC Server handler 2 on default port 45943): State: TIMED_WAITING Blocked count: 0 Waited count: 633 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 107 (IPC Server handler 3 on default port 45943): State: TIMED_WAITING Blocked count: 0 Waited count: 633 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 108 (IPC Server handler 4 on default port 45943): State: TIMED_WAITING Blocked count: 0 Waited count: 633 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 120 (pool-26-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 121 (qtp1040803327-121): State: RUNNABLE Blocked count: 1 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 122 (qtp1040803327-122-acceptor-0@7588b8a-ServerConnector@673672ee{HTTP/1.1, (http/1.1)}{localhost:39721}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 118 (IPC Client (366384994) connection to localhost/127.0.0.1:45939 from jenkins): State: TIMED_WAITING Blocked count: 1578 Waited count: 1575 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Thread 123 (qtp1040803327-123): State: TIMED_WAITING Blocked count: 0 Waited count: 11 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 124 (qtp1040803327-124): State: TIMED_WAITING Blocked count: 0 Waited count: 11 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 119 (IPC Parameter Sending Thread for localhost/127.0.0.1:45939): State: TIMED_WAITING Blocked count: 0 Waited count: 2165 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 125 (Session-HouseKeeper-425819f9-1): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 126 (nioEventLoopGroup-4-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 127 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@49a955f5): State: TIMED_WAITING Blocked count: 0 Waited count: 1264 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 129 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 130 (IPC Server idle connection scanner for port 36851): State: TIMED_WAITING Blocked count: 1 Waited count: 65 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 132 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 127 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 137 (Command processor): State: WAITING Blocked count: 1 Waited count: 329 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@50b25eec Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 138 (BP-1446750426-172.17.0.3-1736312281490 heartbeating to localhost/127.0.0.1:45939): State: TIMED_WAITING Blocked count: 1483 Waited count: 1701 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 139 (pool-29-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 117 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@710dca53): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 131 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 128 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 140 (IPC Server handler 0 on default port 36851): State: TIMED_WAITING Blocked count: 0 Waited count: 632 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 141 (IPC Server handler 1 on default port 36851): State: TIMED_WAITING Blocked count: 0 Waited count: 632 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 142 (IPC Server handler 2 on default port 36851): State: TIMED_WAITING Blocked count: 0 Waited count: 632 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 143 (IPC Server handler 3 on default port 36851): State: TIMED_WAITING Blocked count: 0 Waited count: 632 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 144 (IPC Server handler 4 on default port 36851): State: TIMED_WAITING Blocked count: 0 Waited count: 632 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 154 (pool-36-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 156 (qtp244728962-156): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007ffa3c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 157 (qtp244728962-157-acceptor-0@7ea21e77-ServerConnector@22e0a731{HTTP/1.1, (http/1.1)}{localhost:40719}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 158 (qtp244728962-158): State: TIMED_WAITING Blocked count: 0 Waited count: 11 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 159 (qtp244728962-159): State: TIMED_WAITING Blocked count: 0 Waited count: 11 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 160 (Session-HouseKeeper-562012e4-1): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 162 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data1)): State: TIMED_WAITING Blocked count: 2 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 164 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data2)): State: TIMED_WAITING Blocked count: 4 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 163 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data3)): State: TIMED_WAITING Blocked count: 25 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 165 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data4)): State: TIMED_WAITING Blocked count: 9 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 173 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data3/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 1 Waited count: 3 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 176 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data1/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 2 Waited count: 3 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 175 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data2/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 2 Waited count: 4 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 177 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data4/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 3 Waited count: 3 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 184 (nioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 188 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@1badcc77): State: TIMED_WAITING Blocked count: 0 Waited count: 1263 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 193 (pool-23-thread-1): State: WAITING Blocked count: 0 Waited count: 3 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6d2a9ba3 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 190 (pool-15-thread-1): State: WAITING Blocked count: 0 Waited count: 3 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@c4154b1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 192 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 200 (IPC Server idle connection scanner for port 38935): State: TIMED_WAITING Blocked count: 1 Waited count: 65 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 199 (java.util.concurrent.ThreadPoolExecutor$Worker@3412cacb[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 196 (java.util.concurrent.ThreadPoolExecutor$Worker@17da2a7b[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 202 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 127 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 205 (Command processor): State: WAITING Blocked count: 2 Waited count: 370 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7c245a19 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 206 (BP-1446750426-172.17.0.3-1736312281490 heartbeating to localhost/127.0.0.1:45939): State: TIMED_WAITING Blocked count: 1434 Waited count: 1674 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 207 (pool-46-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 153 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@4ee4c44b): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 201 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 191 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 208 (IPC Server handler 0 on default port 38935): State: TIMED_WAITING Blocked count: 0 Waited count: 634 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 209 (IPC Server handler 1 on default port 38935): State: TIMED_WAITING Blocked count: 0 Waited count: 632 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 210 (IPC Server handler 2 on default port 38935): State: TIMED_WAITING Blocked count: 0 Waited count: 632 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 211 (IPC Server handler 3 on default port 38935): State: TIMED_WAITING Blocked count: 0 Waited count: 632 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 213 (IPC Server handler 4 on default port 38935): State: TIMED_WAITING Blocked count: 0 Waited count: 632 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 215 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data5)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 216 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data6)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 220 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data5/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 2 Waited count: 3 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 222 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data6/current/BP-1446750426-172.17.0.3-1736312281490): State: TIMED_WAITING Blocked count: 2 Waited count: 3 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 226 (pool-33-thread-1): State: WAITING Blocked count: 0 Waited count: 3 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@921ee39 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 229 (java.util.concurrent.ThreadPoolExecutor$Worker@3f6c1323[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 234 (FsDatasetAsyncDiskServiceFixer): State: TIMED_WAITING Blocked count: 0 Waited count: 22 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer.run(HBaseTestingUtility.java:599) Thread 236 (NIOServerCxnFactory.SelectorThread-0): State: RUNNABLE Blocked count: 4 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 237 (NIOServerCxnFactory.SelectorThread-1): State: RUNNABLE Blocked count: 10 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 238 (NIOServerCxnFactory.AcceptThread:localhost/127.0.0.1:59048): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.select(NIOServerCnxnFactory.java:205) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.run(NIOServerCnxnFactory.java:181) Thread 235 (ConnnectionExpirer): State: TIMED_WAITING Blocked count: 0 Waited count: 64 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.NIOServerCnxnFactory$ConnectionExpirerThread.run(NIOServerCnxnFactory.java:554) Thread 239 (SessionTracker): State: TIMED_WAITING Blocked count: 0 Waited count: 316 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.SessionTrackerImpl.run(SessionTrackerImpl.java:163) Thread 240 (SyncThread:0): State: WAITING Blocked count: 22 Waited count: 772 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@51109921 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.SyncRequestProcessor.run(SyncRequestProcessor.java:170) Thread 241 (ProcessThread(sid:0 cport:59048):): State: WAITING Blocked count: 1 Waited count: 876 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5a02804a Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.PrepRequestProcessor.run(PrepRequestProcessor.java:142) Thread 242 (RequestThrottler): State: WAITING Blocked count: 0 Waited count: 900 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5cd4ed15 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.RequestThrottler.run(RequestThrottler.java:147) Thread 243 (NIOWorkerThread-1): State: WAITING Blocked count: 3 Waited count: 138 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 254 (Time-limited test.named-queue-events-pool-0): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@543bccbd Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) app//com.lmax.disruptor.BlockingWaitStrategy.waitFor(BlockingWaitStrategy.java:47) app//com.lmax.disruptor.ProcessingSequenceBarrier.waitFor(ProcessingSequenceBarrier.java:56) app//com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:159) app//com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 255 (HBase-Metrics2-1): State: TIMED_WAITING Blocked count: 0 Waited count: 448 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 256 (RS-EventLoopGroup-1-1): State: RUNNABLE Blocked count: 21 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 257 (Time-limited test-SendThread(127.0.0.1:59048)): State: RUNNABLE Blocked count: 42 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:332) app//org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1289) Thread 258 (Time-limited test-EventThread): State: WAITING Blocked count: 1 Waited count: 59 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@61b2fcae Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:550) Thread 259 (NIOWorkerThread-2): State: WAITING Blocked count: 2 Waited count: 137 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 260 (NIOWorkerThread-3): State: WAITING Blocked count: 2 Waited count: 137 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 261 (NIOWorkerThread-4): State: WAITING Blocked count: 1 Waited count: 137 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 262 (zk-event-processor-pool-0): State: WAITING Blocked count: 38 Waited count: 91 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5c993aff Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 263 (NIOWorkerThread-5): State: WAITING Blocked count: 3 Waited count: 138 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 264 (NIOWorkerThread-6): State: WAITING Blocked count: 2 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 265 (NIOWorkerThread-7): State: WAITING Blocked count: 3 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 266 (NIOWorkerThread-8): State: WAITING Blocked count: 1 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 267 (NIOWorkerThread-9): State: WAITING Blocked count: 3 Waited count: 137 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 268 (NIOWorkerThread-10): State: WAITING Blocked count: 2 Waited count: 137 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 269 (NIOWorkerThread-11): State: WAITING Blocked count: 1 Waited count: 137 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 270 (NIOWorkerThread-12): State: WAITING Blocked count: 0 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 271 (NIOWorkerThread-13): State: WAITING Blocked count: 2 Waited count: 137 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 272 (NIOWorkerThread-14): State: WAITING Blocked count: 2 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 273 (NIOWorkerThread-15): State: WAITING Blocked count: 4 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 274 (NIOWorkerThread-16): State: WAITING Blocked count: 2 Waited count: 137 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b7843b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 275 (RpcServer.default.FPBQ.Fifo.handler=0,queue=0,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@564d5bf5 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 276 (RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=42195): State: WAITING Blocked count: 109 Waited count: 415 Waiting on java.util.concurrent.Semaphore$NonfairSync@643f078a Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 277 (RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42195): State: WAITING Blocked count: 203 Waited count: 823 Waiting on java.util.concurrent.Semaphore$NonfairSync@129ed02f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 278 (RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42195): State: WAITING Blocked count: 38 Waited count: 9161 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7c54e41 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 279 (RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@3b6b3fbf Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 280 (RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@3b6b3fbf Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 281 (RpcServer.replication.FPBQ.Fifo.handler=0,queue=0,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@27f47c0f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 282 (RpcServer.replication.FPBQ.Fifo.handler=1,queue=0,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@438fcf76 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 283 (RpcServer.replication.FPBQ.Fifo.handler=2,queue=0,port=42195): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@6e953600 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 284 (RpcServer.metaPriority.FPBQ.Fifo.handler=0,queue=0,port=42195): State: WAITING Blocked count: 0 Waited count: 7 Waiting on java.util.concurrent.Semaphore$NonfairSync@ce17aaa Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 288 (RS-EventLoopGroup-3-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 308 (RS-EventLoopGroup-4-1): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 328 (RS-EventLoopGroup-5-1): State: RUNNABLE Blocked count: 68 Waited count: 3 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 286 (M:0;f28094b72178:42195): State: TIMED_WAITING Blocked count: 8 Waited count: 3269 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.FutureTask.awaitDone(FutureTask.java:444) java.base@17.0.11/java.util.concurrent.FutureTask.get(FutureTask.java:203) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.shutdown(AbstractFSWAL.java:1011) app//org.apache.hadoop.hbase.wal.AbstractFSWALProvider.shutdown(AbstractFSWALProvider.java:184) app//org.apache.hadoop.hbase.wal.WALFactory.shutdown(WALFactory.java:272) app//org.apache.hadoop.hbase.master.region.MasterRegion.shutdownWAL(MasterRegion.java:140) app//org.apache.hadoop.hbase.master.region.MasterRegion.close(MasterRegion.java:206) app//org.apache.hadoop.hbase.master.HMaster.stopServiceThreads(HMaster.java:1769) app//org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:1285) app//org.apache.hadoop.hbase.master.HMaster.run(HMaster.java:614) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 349 (Monitor thread for TaskMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 63 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.monitoring.TaskMonitor$MonitorRunnable.run(TaskMonitor.java:325) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 351 (master/f28094b72178:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 3 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 353 (master/f28094b72178:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 3 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 355 (org.apache.hadoop.hdfs.PeerCache@7594f538): State: TIMED_WAITING Blocked count: 0 Waited count: 210 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.PeerCache.run(PeerCache.java:253) app//org.apache.hadoop.hdfs.PeerCache.access$000(PeerCache.java:46) app//org.apache.hadoop.hdfs.PeerCache$1.run(PeerCache.java:124) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 374 (master:store-WAL-Roller): State: TIMED_WAITING Blocked count: 0 Waited count: 6253 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hbase.wal.AbstractWALRoller.run(AbstractWALRoller.java:179) Thread 391 (RS-EventLoopGroup-5-2): State: RUNNABLE Blocked count: 53 Waited count: 4 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 392 (RS-EventLoopGroup-5-3): State: RUNNABLE Blocked count: 37 Waited count: 4 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 404 (Idle-Rpc-Conn-Sweeper-pool-0): State: WAITING Blocked count: 0 Waited count: 67 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@75e9f460 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 415 (SnapshotHandlerChoreCleaner): State: TIMED_WAITING Blocked count: 0 Waited count: 63 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 403 (RpcClient-timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 62500 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 424 (RS-EventLoopGroup-1-2): State: RUNNABLE Blocked count: 50 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 425 (RS-EventLoopGroup-1-3): State: RUNNABLE Blocked count: 34 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 440 (RegionServerTracker-0): State: WAITING Blocked count: 8 Waited count: 11 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4814fa0 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 471 (regionserver/f28094b72178:0.procedureResultReporter): State: WAITING Blocked count: 13 Waited count: 27 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@21db0ca7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 476 (regionserver/f28094b72178:0.procedureResultReporter): State: WAITING Blocked count: 19 Waited count: 39 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1b2ba608 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 467 (regionserver/f28094b72178:0.procedureResultReporter): State: WAITING Blocked count: 9 Waited count: 19 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7f063557 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 512 (RS-EventLoopGroup-4-2): State: RUNNABLE Blocked count: 2 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 523 (RS-EventLoopGroup-4-3): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 525 (region-location-0): State: WAITING Blocked count: 7 Waited count: 12 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@71749958 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 551 (Async-Client-Retry-Timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 62234 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 565 (RPCClient-NioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 8 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 604 (region-location-1): State: WAITING Blocked count: 6 Waited count: 11 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@71749958 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 605 (region-location-2): State: WAITING Blocked count: 4 Waited count: 9 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@71749958 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 606 (region-location-3): State: WAITING Blocked count: 3 Waited count: 8 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@71749958 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1007 (MutableQuantiles-0): State: TIMED_WAITING Blocked count: 0 Waited count: 438 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1071 (RPCClient-NioEventLoopGroup-6-2): State: RUNNABLE Blocked count: 26 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1111 (zk-permission-watcher-pool-0): State: WAITING Blocked count: 61 Waited count: 92 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@8cc91a Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1119 (RPCClient-NioEventLoopGroup-6-3): State: RUNNABLE Blocked count: 5 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1567 (Container metrics unregistration): State: WAITING Blocked count: 10 Waited count: 33 Waiting on java.util.TaskQueue@55e73e0a Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 2046 (RS-EventLoopGroup-3-2): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 2058 (RPCClient-NioEventLoopGroup-6-4): State: RUNNABLE Blocked count: 2 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 2059 (RS-EventLoopGroup-3-3): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 2420 (region-location-4): State: WAITING Blocked count: 2 Waited count: 7 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@71749958 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5232 (RPCClient-NioEventLoopGroup-6-5): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5233 (RPCClient-NioEventLoopGroup-6-6): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5234 (RPCClient-NioEventLoopGroup-6-7): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9459 (AsyncFSWAL-1-hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/MasterData-prefix:f28094b72178,42195,1736312287279): State: WAITING Blocked count: 0 Waited count: 2 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@452b5d66 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9463 (process reaper): State: TIMED_WAITING Blocked count: 0 Waited count: 6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:401) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9470 (Timer for 'JobHistoryServer' metrics system): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 9471 (WAL-Shutdown-0): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.awaitTermination(ThreadPoolExecutor.java:1464) app//org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.doShutdown(AsyncFSWAL.java:793) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL$2.call(AbstractFSWAL.java:995) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL$2.call(AbstractFSWAL.java:990) java.base@17.0.11/java.util.concurrent.FutureTask.run(FutureTask.java:264) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9472 (Close-WAL-Writer-0): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:166) app//org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) app//org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.recoverLease(AsyncFSWAL.java:722) app//org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.lambda$closeWriter$5(AsyncFSWAL.java:735) app//org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL$$Lambda$1137/0x00007ffa3d17b6f8.run(Unknown Source) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) 2025-01-08T05:08:40,268 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(233): attempt=1 on file=hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/MasterData/WALs/f28094b72178,42195,1736312287279/f28094b72178%2C42195%2C1736312287279.1736312289851 after 4001ms java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.recoverLease(DFSClient.java:946) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:317) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:314) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.recoverLease(DistributedFileSystem.java:329) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverLease(RecoverLeaseFSUtils.java:222) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:155) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.recoverLease(AsyncFSWAL.java:722) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.lambda$closeWriter$5(AsyncFSWAL.java:735) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2025-01-08T05:08:41,264 ERROR [WAL-Shutdown-0 {}] wal.AsyncFSWAL(794): We have waited 5 seconds but the close of async writer doesn't complete.Please check the status of underlying filesystem or increase the wait time by the config "hbase.wal.async.wait.on.shutdown.seconds" 2025-01-08T05:08:41,265 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2025-01-08T05:08:41,265 INFO [M:0;f28094b72178:42195 {}] flush.MasterFlushTableProcedureManager(91): stop: server shutting down. 2025-01-08T05:08:41,265 INFO [M:0;f28094b72178:42195 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.3:42195 2025-01-08T05:08:41,267 DEBUG [M:0;f28094b72178:42195 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/rs/f28094b72178,42195,1736312287279 already deleted, retry=false 2025-01-08T05:08:41,270 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45939/user/jenkins/test-data/622411d3-4d14-7a1a-499c-7e3707b8df49/MasterData/WALs/f28094b72178,42195,1736312287279/f28094b72178%2C42195%2C1736312287279.1736312289851 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.recoverLease(AsyncFSWAL.java:722) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.lambda$closeWriter$5(AsyncFSWAL.java:735) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 12 more 2025-01-08T05:08:41,368 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2025-01-08T05:08:41,368 INFO [M:0;f28094b72178:42195 {}] regionserver.HRegionServer(1307): Exiting; stopping=f28094b72178,42195,1736312287279; zookeeper connection closed. 2025-01-08T05:08:41,368 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42195-0x1000ae1399f0000, quorum=127.0.0.1:59048, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2025-01-08T05:08:41,372 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@85b88e0{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2025-01-08T05:08:41,372 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@22e0a731{HTTP/1.1, (http/1.1)}{localhost:0} 2025-01-08T05:08:41,373 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2025-01-08T05:08:41,373 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3564bcd4{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2025-01-08T05:08:41,373 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@2f88ae23{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop.log.dir/,STOPPED} 2025-01-08T05:08:41,375 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2025-01-08T05:08:41,375 WARN [BP-1446750426-172.17.0.3-1736312281490 heartbeating to localhost/127.0.0.1:45939 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2025-01-08T05:08:41,375 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2025-01-08T05:08:41,375 WARN [BP-1446750426-172.17.0.3-1736312281490 heartbeating to localhost/127.0.0.1:45939 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1446750426-172.17.0.3-1736312281490 (Datanode Uuid ec76c1ae-a7ea-45af-8315-9b1b87ea684c) service to localhost/127.0.0.1:45939 2025-01-08T05:08:41,376 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data5/current/BP-1446750426-172.17.0.3-1736312281490 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2025-01-08T05:08:41,376 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data6/current/BP-1446750426-172.17.0.3-1736312281490 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2025-01-08T05:08:41,377 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2025-01-08T05:08:41,379 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@13469af5{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2025-01-08T05:08:41,380 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@673672ee{HTTP/1.1, (http/1.1)}{localhost:0} 2025-01-08T05:08:41,380 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2025-01-08T05:08:41,380 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@53b7d93e{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2025-01-08T05:08:41,380 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@4cbe4b5e{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop.log.dir/,STOPPED} 2025-01-08T05:08:41,381 WARN [BP-1446750426-172.17.0.3-1736312281490 heartbeating to localhost/127.0.0.1:45939 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2025-01-08T05:08:41,381 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2025-01-08T05:08:41,381 WARN [BP-1446750426-172.17.0.3-1736312281490 heartbeating to localhost/127.0.0.1:45939 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1446750426-172.17.0.3-1736312281490 (Datanode Uuid e8db6b00-be0c-4265-bbbd-8f000659b075) service to localhost/127.0.0.1:45939 2025-01-08T05:08:41,381 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2025-01-08T05:08:41,382 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data3/current/BP-1446750426-172.17.0.3-1736312281490 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2025-01-08T05:08:41,382 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data4/current/BP-1446750426-172.17.0.3-1736312281490 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2025-01-08T05:08:41,383 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2025-01-08T05:08:41,385 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@d799fa8{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2025-01-08T05:08:41,385 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@7455429{HTTP/1.1, (http/1.1)}{localhost:0} 2025-01-08T05:08:41,385 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2025-01-08T05:08:41,386 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3c53997b{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2025-01-08T05:08:41,386 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@5ce17cd2{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop.log.dir/,STOPPED} 2025-01-08T05:08:41,388 WARN [BP-1446750426-172.17.0.3-1736312281490 heartbeating to localhost/127.0.0.1:45939 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2025-01-08T05:08:41,388 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2025-01-08T05:08:41,388 WARN [BP-1446750426-172.17.0.3-1736312281490 heartbeating to localhost/127.0.0.1:45939 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1446750426-172.17.0.3-1736312281490 (Datanode Uuid 952bc167-5e99-4eb6-aae6-febac4c94286) service to localhost/127.0.0.1:45939 2025-01-08T05:08:41,388 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2025-01-08T05:08:41,389 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data1/current/BP-1446750426-172.17.0.3-1736312281490 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2025-01-08T05:08:41,390 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/cluster_33570ea1-d81e-112d-1be2-03bb0d06aeab/dfs/data/data2/current/BP-1446750426-172.17.0.3-1736312281490 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2025-01-08T05:08:41,390 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2025-01-08T05:08:41,399 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@27047d40{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2025-01-08T05:08:41,400 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@61d1e83b{HTTP/1.1, (http/1.1)}{localhost:0} 2025-01-08T05:08:41,400 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2025-01-08T05:08:41,400 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@15c67ac2{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2025-01-08T05:08:41,400 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@19c1c4e1{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/5feb2ffc-5d84-eb73-053e-fe5b5c78555d/hadoop.log.dir/,STOPPED} 2025-01-08T05:08:41,417 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(345): Shutdown MiniZK cluster with all ZK servers 2025-01-08T05:08:41,757 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1347): Minicluster is down