2025-01-08 10:44:58,746 main DEBUG Apache Log4j Core 2.17.2 initializing configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@3de8f619 2025-01-08 10:44:58,767 main DEBUG Took 0.018238 seconds to load 1 plugins from package org.apache.hadoop.hbase.logging 2025-01-08 10:44:58,768 main DEBUG PluginManager 'Core' found 129 plugins 2025-01-08 10:44:58,768 main DEBUG PluginManager 'Level' found 0 plugins 2025-01-08 10:44:58,772 main DEBUG PluginManager 'Lookup' found 16 plugins 2025-01-08 10:44:58,775 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 10:44:58,785 main DEBUG PluginManager 'TypeConverter' found 26 plugins 2025-01-08 10:44:58,803 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.util.MBeans", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 10:44:58,806 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 10:44:58,807 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.logging.TestJul2Slf4j", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 10:44:58,807 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 10:44:58,808 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.zookeeper", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 10:44:58,808 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 10:44:58,810 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSinkAdapter", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 10:44:58,810 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 10:44:58,811 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSystemImpl", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 10:44:58,811 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 10:44:58,812 main DEBUG LoggerConfig$Builder(additivity="false", level="WARN", levelAndRefs="null", name="org.apache.directory", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 10:44:58,813 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 10:44:58,814 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.ipc.FailedServers", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 10:44:58,814 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 10:44:58,814 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsConfig", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 10:44:58,815 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 10:44:58,815 main DEBUG LoggerConfig$Builder(additivity="null", level="INFO", levelAndRefs="null", name="org.apache.hadoop.hbase.ScheduledChore", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 10:44:58,816 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 10:44:58,816 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.regionserver.RSRpcServices", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 10:44:58,818 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 10:44:58,818 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 10:44:58,819 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 10:44:58,820 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 10:44:58,820 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 10:44:58,822 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hbase.thirdparty.io.netty.channel", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 10:44:58,822 main DEBUG Building Plugin[name=root, class=org.apache.logging.log4j.core.config.LoggerConfig$RootLogger]. 2025-01-08 10:44:58,826 main DEBUG LoggerConfig$RootLogger$Builder(additivity="null", level="null", levelAndRefs="INFO,Console", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 10:44:58,828 main DEBUG Building Plugin[name=loggers, class=org.apache.logging.log4j.core.config.LoggersPlugin]. 2025-01-08 10:44:58,838 main DEBUG createLoggers(={org.apache.hadoop.metrics2.util.MBeans, org.apache.hadoop.hbase.logging.TestJul2Slf4j, org.apache.zookeeper, org.apache.hadoop.metrics2.impl.MetricsSinkAdapter, org.apache.hadoop.metrics2.impl.MetricsSystemImpl, org.apache.directory, org.apache.hadoop.hbase.ipc.FailedServers, org.apache.hadoop.metrics2.impl.MetricsConfig, org.apache.hadoop.hbase.ScheduledChore, org.apache.hadoop.hbase.regionserver.RSRpcServices, org.apache.hadoop, org.apache.hadoop.hbase, org.apache.hbase.thirdparty.io.netty.channel, root}) 2025-01-08 10:44:58,839 main DEBUG Building Plugin[name=layout, class=org.apache.logging.log4j.core.layout.PatternLayout]. 2025-01-08 10:44:58,841 main DEBUG PatternLayout$Builder(pattern="%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n", PatternSelector=null, Configuration(PropertiesConfig), Replace=null, charset="null", alwaysWriteExceptions="null", disableAnsi="null", noConsoleNoAnsi="null", header="null", footer="null") 2025-01-08 10:44:58,841 main DEBUG PluginManager 'Converter' found 47 plugins 2025-01-08 10:44:58,854 main DEBUG Building Plugin[name=appender, class=org.apache.hadoop.hbase.logging.HBaseTestAppender]. 2025-01-08 10:44:58,857 main DEBUG HBaseTestAppender$Builder(target="SYSTEM_ERR", maxSize="1G", bufferedIo="null", bufferSize="null", immediateFlush="null", ignoreExceptions="null", PatternLayout(%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n), name="Console", Configuration(PropertiesConfig), Filter=null, ={}) 2025-01-08 10:44:58,867 main DEBUG Starting HBaseTestOutputStreamManager SYSTEM_ERR 2025-01-08 10:44:58,867 main DEBUG Building Plugin[name=appenders, class=org.apache.logging.log4j.core.config.AppendersPlugin]. 2025-01-08 10:44:58,868 main DEBUG createAppenders(={Console}) 2025-01-08 10:44:58,869 main DEBUG Configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@3de8f619 initialized 2025-01-08 10:44:58,869 main DEBUG Starting configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@3de8f619 2025-01-08 10:44:58,870 main DEBUG Started configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@3de8f619 OK. 2025-01-08 10:44:58,873 main DEBUG Shutting down OutputStreamManager SYSTEM_OUT.false.false-1 2025-01-08 10:44:58,873 main DEBUG OutputStream closed 2025-01-08 10:44:58,876 main DEBUG Shut down OutputStreamManager SYSTEM_OUT.false.false-1, all resources released: true 2025-01-08 10:44:58,879 main DEBUG Appender DefaultConsole-1 stopped with status true 2025-01-08 10:44:58,880 main DEBUG Stopped org.apache.logging.log4j.core.config.DefaultConfiguration@58695725 OK 2025-01-08 10:44:59,039 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6 2025-01-08 10:44:59,045 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=StatusLogger 2025-01-08 10:44:59,047 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=ContextSelector 2025-01-08 10:44:59,048 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name= 2025-01-08 10:44:59,049 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.directory 2025-01-08 10:44:59,050 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSinkAdapter 2025-01-08 10:44:59,050 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.zookeeper 2025-01-08 10:44:59,051 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.logging.TestJul2Slf4j 2025-01-08 10:44:59,051 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSystemImpl 2025-01-08 10:44:59,051 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.util.MBeans 2025-01-08 10:44:59,052 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase 2025-01-08 10:44:59,052 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop 2025-01-08 10:44:59,053 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ipc.FailedServers 2025-01-08 10:44:59,053 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.regionserver.RSRpcServices 2025-01-08 10:44:59,054 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsConfig 2025-01-08 10:44:59,054 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hbase.thirdparty.io.netty.channel 2025-01-08 10:44:59,062 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ScheduledChore 2025-01-08 10:44:59,063 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Appenders,name=Console 2025-01-08 10:44:59,066 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2025-01-08 10:44:59,066 main DEBUG Reconfiguration complete for context[name=1dbd16a6] at URI jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-logging/target/hbase-logging-2.7.0-SNAPSHOT-tests.jar!/log4j2.properties (org.apache.logging.log4j.core.LoggerContext@49ff7d8c) with optional ClassLoader: null 2025-01-08 10:44:59,067 main DEBUG Shutdown hook enabled. Registering a new one. 2025-01-08 10:44:59,068 main DEBUG LoggerContext[name=1dbd16a6, org.apache.logging.log4j.core.LoggerContext@49ff7d8c] started OK. 2025-01-08T10:44:59,086 INFO [main {}] hbase.HBaseClassTestRule(94): Test class org.apache.hadoop.hbase.snapshot.TestExportSnapshot timeout: 13 mins 2025-01-08 10:44:59,090 main DEBUG AsyncLogger.ThreadNameStrategy=UNCACHED (user specified null, default is UNCACHED) 2025-01-08 10:44:59,090 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2025-01-08T10:44:59,655 DEBUG [main {}] hbase.HBaseTestingUtility(348): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193 2025-01-08T10:44:59,656 INFO [main {}] hbase.HBaseClassTestRule(94): Test class org.apache.hadoop.hbase.snapshot.TestMobExportSnapshot timeout: 13 mins 2025-01-08T10:44:59,656 INFO [main {}] hbase.HBaseClassTestRule(94): Test class org.apache.hadoop.hbase.snapshot.TestMobSecureExportSnapshot timeout: 13 mins 2025-01-08T10:44:59,747 WARN [Time-limited test {}] util.NativeCodeLoader(60): Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2025-01-08T10:45:00,126 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1126): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2025-01-08T10:45:00,127 INFO [Time-limited test {}] hbase.HBaseZKTestingUtility(82): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204, deleteOnExit=true 2025-01-08T10:45:00,131 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1140): STARTING DFS 2025-01-08T10:45:00,132 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/test.cache.data in system properties and HBase conf 2025-01-08T10:45:00,133 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop.tmp.dir in system properties and HBase conf 2025-01-08T10:45:00,134 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop.log.dir in system properties and HBase conf 2025-01-08T10:45:00,134 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/mapreduce.cluster.local.dir in system properties and HBase conf 2025-01-08T10:45:00,135 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/mapreduce.cluster.temp.dir in system properties and HBase conf 2025-01-08T10:45:00,135 INFO [Time-limited test {}] hbase.HBaseTestingUtility(811): read short circuit is OFF 2025-01-08T10:45:00,313 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2025-01-08T10:45:00,321 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2025-01-08T10:45:00,335 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2025-01-08T10:45:00,341 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/yarn.nodemanager.log-dirs in system properties and HBase conf 2025-01-08T10:45:00,341 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2025-01-08T10:45:00,342 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2025-01-08T10:45:00,343 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2025-01-08T10:45:00,344 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2025-01-08T10:45:00,345 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/dfs.journalnode.edits.dir in system properties and HBase conf 2025-01-08T10:45:00,345 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2025-01-08T10:45:00,346 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/nfs.dump.dir in system properties and HBase conf 2025-01-08T10:45:00,347 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/java.io.tmpdir in system properties and HBase conf 2025-01-08T10:45:00,348 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/dfs.journalnode.edits.dir in system properties and HBase conf 2025-01-08T10:45:00,349 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2025-01-08T10:45:00,349 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2025-01-08T10:45:02,244 WARN [Time-limited test {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties 2025-01-08T10:45:02,347 INFO [Time-limited test {}] log.Log(170): Logging initialized @4584ms to org.eclipse.jetty.util.log.Slf4jLog 2025-01-08T10:45:02,468 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T10:45:02,564 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2025-01-08T10:45:02,637 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2025-01-08T10:45:02,637 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2025-01-08T10:45:02,639 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2025-01-08T10:45:02,659 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T10:45:02,673 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@d3238fd{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop.log.dir/,AVAILABLE} 2025-01-08T10:45:02,674 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@1ba6b2fb{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2025-01-08T10:45:02,937 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@146536f9{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/java.io.tmpdir/jetty-localhost-42647-hadoop-hdfs-3_4_1-tests_jar-_-any-9586633205635398905/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2025-01-08T10:45:02,948 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@67ec1049{HTTP/1.1, (http/1.1)}{localhost:42647} 2025-01-08T10:45:02,949 INFO [Time-limited test {}] server.Server(415): Started @5188ms 2025-01-08T10:45:03,463 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T10:45:03,475 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2025-01-08T10:45:03,498 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2025-01-08T10:45:03,498 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2025-01-08T10:45:03,499 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2025-01-08T10:45:03,502 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@617ad2bd{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop.log.dir/,AVAILABLE} 2025-01-08T10:45:03,503 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@666781c4{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2025-01-08T10:45:03,694 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@5a59e1d7{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/java.io.tmpdir/jetty-localhost-34891-hadoop-hdfs-3_4_1-tests_jar-_-any-9059764945362794805/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2025-01-08T10:45:03,699 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@3c5cb06a{HTTP/1.1, (http/1.1)}{localhost:34891} 2025-01-08T10:45:03,700 INFO [Time-limited test {}] server.Server(415): Started @5939ms 2025-01-08T10:45:03,801 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2025-01-08T10:45:04,208 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T10:45:04,229 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2025-01-08T10:45:04,253 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2025-01-08T10:45:04,253 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2025-01-08T10:45:04,254 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2025-01-08T10:45:04,256 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@66733fff{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop.log.dir/,AVAILABLE} 2025-01-08T10:45:04,257 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6c51149d{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2025-01-08T10:45:04,496 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@3abf71c7{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/java.io.tmpdir/jetty-localhost-42931-hadoop-hdfs-3_4_1-tests_jar-_-any-6044404802374358961/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2025-01-08T10:45:04,498 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@27b3056c{HTTP/1.1, (http/1.1)}{localhost:42931} 2025-01-08T10:45:04,499 INFO [Time-limited test {}] server.Server(415): Started @6738ms 2025-01-08T10:45:04,506 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2025-01-08T10:45:04,740 WARN [Thread-93 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data1/current/BP-22138141-172.17.0.2-1736333101274/current, will proceed with Du for space computation calculation, 2025-01-08T10:45:04,756 WARN [Thread-96 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data2/current/BP-22138141-172.17.0.2-1736333101274/current, will proceed with Du for space computation calculation, 2025-01-08T10:45:04,780 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T10:45:04,790 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2025-01-08T10:45:04,808 WARN [Thread-107 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data3/current/BP-22138141-172.17.0.2-1736333101274/current, will proceed with Du for space computation calculation, 2025-01-08T10:45:04,815 WARN [Thread-108 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data4/current/BP-22138141-172.17.0.2-1736333101274/current, will proceed with Du for space computation calculation, 2025-01-08T10:45:04,821 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2025-01-08T10:45:04,821 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2025-01-08T10:45:04,822 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2025-01-08T10:45:04,835 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@29173d0{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop.log.dir/,AVAILABLE} 2025-01-08T10:45:04,839 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@1e8b912e{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2025-01-08T10:45:04,874 WARN [Thread-58 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2025-01-08T10:45:04,920 WARN [Thread-83 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2025-01-08T10:45:04,985 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xffcea3966f5fb756 with lease ID 0xe506f1aed76802ff: Processing first storage report for DS-96de2396-f7dd-4dd7-805c-c34c70b41204 from datanode DatanodeRegistration(127.0.0.1:43487, datanodeUuid=2039b132-af90-4efa-b0f3-2607003b309d, infoPort=34377, infoSecurePort=0, ipcPort=45989, storageInfo=lv=-57;cid=testClusterID;nsid=1138699459;c=1736333101274) 2025-01-08T10:45:04,986 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xffcea3966f5fb756 with lease ID 0xe506f1aed76802ff: from storage DS-96de2396-f7dd-4dd7-805c-c34c70b41204 node DatanodeRegistration(127.0.0.1:43487, datanodeUuid=2039b132-af90-4efa-b0f3-2607003b309d, infoPort=34377, infoSecurePort=0, ipcPort=45989, storageInfo=lv=-57;cid=testClusterID;nsid=1138699459;c=1736333101274), blocks: 0, hasStaleStorage: true, processing time: 2 msecs, invalidatedBlocks: 0 2025-01-08T10:45:04,987 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x3938d050ceed1474 with lease ID 0xe506f1aed76802fe: Processing first storage report for DS-687c546a-3440-42d4-b346-ee987087ce43 from datanode DatanodeRegistration(127.0.0.1:44473, datanodeUuid=b55dd1b9-f145-4b75-892c-153c95fea6e1, infoPort=40217, infoSecurePort=0, ipcPort=37919, storageInfo=lv=-57;cid=testClusterID;nsid=1138699459;c=1736333101274) 2025-01-08T10:45:04,987 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x3938d050ceed1474 with lease ID 0xe506f1aed76802fe: from storage DS-687c546a-3440-42d4-b346-ee987087ce43 node DatanodeRegistration(127.0.0.1:44473, datanodeUuid=b55dd1b9-f145-4b75-892c-153c95fea6e1, infoPort=40217, infoSecurePort=0, ipcPort=37919, storageInfo=lv=-57;cid=testClusterID;nsid=1138699459;c=1736333101274), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2025-01-08T10:45:04,987 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xffcea3966f5fb756 with lease ID 0xe506f1aed76802ff: Processing first storage report for DS-3a6862aa-5072-4346-a310-a3fc8b9fc69c from datanode DatanodeRegistration(127.0.0.1:43487, datanodeUuid=2039b132-af90-4efa-b0f3-2607003b309d, infoPort=34377, infoSecurePort=0, ipcPort=45989, storageInfo=lv=-57;cid=testClusterID;nsid=1138699459;c=1736333101274) 2025-01-08T10:45:04,988 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xffcea3966f5fb756 with lease ID 0xe506f1aed76802ff: from storage DS-3a6862aa-5072-4346-a310-a3fc8b9fc69c node DatanodeRegistration(127.0.0.1:43487, datanodeUuid=2039b132-af90-4efa-b0f3-2607003b309d, infoPort=34377, infoSecurePort=0, ipcPort=45989, storageInfo=lv=-57;cid=testClusterID;nsid=1138699459;c=1736333101274), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2025-01-08T10:45:04,988 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x3938d050ceed1474 with lease ID 0xe506f1aed76802fe: Processing first storage report for DS-9263468e-af3e-4dcf-ab9d-30bb139a42d4 from datanode DatanodeRegistration(127.0.0.1:44473, datanodeUuid=b55dd1b9-f145-4b75-892c-153c95fea6e1, infoPort=40217, infoSecurePort=0, ipcPort=37919, storageInfo=lv=-57;cid=testClusterID;nsid=1138699459;c=1736333101274) 2025-01-08T10:45:04,988 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x3938d050ceed1474 with lease ID 0xe506f1aed76802fe: from storage DS-9263468e-af3e-4dcf-ab9d-30bb139a42d4 node DatanodeRegistration(127.0.0.1:44473, datanodeUuid=b55dd1b9-f145-4b75-892c-153c95fea6e1, infoPort=40217, infoSecurePort=0, ipcPort=37919, storageInfo=lv=-57;cid=testClusterID;nsid=1138699459;c=1736333101274), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2025-01-08T10:45:05,015 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@6cc6a12e{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/java.io.tmpdir/jetty-localhost-40601-hadoop-hdfs-3_4_1-tests_jar-_-any-9037022750161450365/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2025-01-08T10:45:05,016 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@3362f87c{HTTP/1.1, (http/1.1)}{localhost:40601} 2025-01-08T10:45:05,016 INFO [Time-limited test {}] server.Server(415): Started @7256ms 2025-01-08T10:45:05,020 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2025-01-08T10:45:05,286 WARN [Thread-139 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data5/current/BP-22138141-172.17.0.2-1736333101274/current, will proceed with Du for space computation calculation, 2025-01-08T10:45:05,287 WARN [Thread-140 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data6/current/BP-22138141-172.17.0.2-1736333101274/current, will proceed with Du for space computation calculation, 2025-01-08T10:45:05,371 WARN [Thread-129 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2025-01-08T10:45:05,386 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x87ec3d514a7621c2 with lease ID 0xe506f1aed7680300: Processing first storage report for DS-5709618f-16cd-4409-9a05-d0991502eaf3 from datanode DatanodeRegistration(127.0.0.1:45285, datanodeUuid=eceeeec2-0c7e-4597-a09d-1a300452c646, infoPort=41419, infoSecurePort=0, ipcPort=45913, storageInfo=lv=-57;cid=testClusterID;nsid=1138699459;c=1736333101274) 2025-01-08T10:45:05,386 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x87ec3d514a7621c2 with lease ID 0xe506f1aed7680300: from storage DS-5709618f-16cd-4409-9a05-d0991502eaf3 node DatanodeRegistration(127.0.0.1:45285, datanodeUuid=eceeeec2-0c7e-4597-a09d-1a300452c646, infoPort=41419, infoSecurePort=0, ipcPort=45913, storageInfo=lv=-57;cid=testClusterID;nsid=1138699459;c=1736333101274), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2025-01-08T10:45:05,387 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x87ec3d514a7621c2 with lease ID 0xe506f1aed7680300: Processing first storage report for DS-b3c6b5ad-5d77-44f8-84b6-11118b54404c from datanode DatanodeRegistration(127.0.0.1:45285, datanodeUuid=eceeeec2-0c7e-4597-a09d-1a300452c646, infoPort=41419, infoSecurePort=0, ipcPort=45913, storageInfo=lv=-57;cid=testClusterID;nsid=1138699459;c=1736333101274) 2025-01-08T10:45:05,387 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x87ec3d514a7621c2 with lease ID 0xe506f1aed7680300: from storage DS-b3c6b5ad-5d77-44f8-84b6-11118b54404c node DatanodeRegistration(127.0.0.1:45285, datanodeUuid=eceeeec2-0c7e-4597-a09d-1a300452c646, infoPort=41419, infoSecurePort=0, ipcPort=45913, storageInfo=lv=-57;cid=testClusterID;nsid=1138699459;c=1736333101274), blocks: 0, hasStaleStorage: false, processing time: 1 msecs, invalidatedBlocks: 0 2025-01-08T10:45:05,709 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(703): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193 2025-01-08T10:45:05,844 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(259): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/zookeeper_0, clientPort=52367, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2025-01-08T10:45:05,874 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(284): Started MiniZooKeeperCluster and ran 'stat' on client port=52367 2025-01-08T10:45:05,889 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T10:45:05,893 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T10:45:06,345 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741825_1001 (size=7) 2025-01-08T10:45:06,347 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741825_1001 (size=7) 2025-01-08T10:45:06,347 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741825_1001 (size=7) 2025-01-08T10:45:06,754 INFO [Time-limited test {}] util.FSUtils(490): Created version file at hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f with version=8 2025-01-08T10:45:06,755 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1460): Setting hbase.fs.tmp.dir to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/hbase-staging 2025-01-08T10:45:06,967 DEBUG [Time-limited test {}] channel.MultithreadEventLoopGroup(44): -Dio.netty.eventLoopThreads: 16 2025-01-08T10:45:07,305 INFO [Time-limited test {}] client.ConnectionUtils(129): master/67f42ee69b4e:0 server-side Connection retries=45 2025-01-08T10:45:07,332 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2025-01-08T10:45:07,333 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2025-01-08T10:45:07,333 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2025-01-08T10:45:07,333 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2025-01-08T10:45:07,334 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2025-01-08T10:45:07,522 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2025-01-08T10:45:07,600 INFO [Time-limited test {}] metrics.MetricRegistriesLoader(60): Loaded MetricRegistries class org.apache.hadoop.hbase.metrics.impl.MetricRegistriesImpl 2025-01-08T10:45:07,612 DEBUG [Time-limited test {}] util.ClassSize(228): Using Unsafe to estimate memory layout 2025-01-08T10:45:07,617 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2025-01-08T10:45:07,648 DEBUG [Time-limited test {}] channel.DefaultChannelId(84): -Dio.netty.processId: 32206 (auto-detected) 2025-01-08T10:45:07,649 DEBUG [Time-limited test {}] channel.DefaultChannelId(106): -Dio.netty.machineId: 02:42:ac:ff:fe:11:00:02 (auto-detected) 2025-01-08T10:45:07,671 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:34551 2025-01-08T10:45:07,681 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T10:45:07,684 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T10:45:07,697 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=master:34551 connecting to ZooKeeper ensemble=127.0.0.1:52367 2025-01-08T10:45:07,746 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:345510x0, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2025-01-08T10:45:07,777 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:34551-0x1000c1ee94b0000 connected 2025-01-08T10:45:07,868 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2025-01-08T10:45:07,872 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2025-01-08T10:45:07,897 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2025-01-08T10:45:07,902 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=34551 2025-01-08T10:45:07,903 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=34551 2025-01-08T10:45:07,904 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=34551 2025-01-08T10:45:07,905 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=34551 2025-01-08T10:45:07,906 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=34551 2025-01-08T10:45:07,920 INFO [Time-limited test {}] master.HMaster(499): hbase.rootdir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f, hbase.cluster.distributed=false 2025-01-08T10:45:08,041 INFO [Time-limited test {}] client.ConnectionUtils(129): regionserver/67f42ee69b4e:0 server-side Connection retries=45 2025-01-08T10:45:08,041 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2025-01-08T10:45:08,041 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2025-01-08T10:45:08,042 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2025-01-08T10:45:08,042 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2025-01-08T10:45:08,042 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2025-01-08T10:45:08,045 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2025-01-08T10:45:08,050 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2025-01-08T10:45:08,068 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:45003 2025-01-08T10:45:08,071 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2025-01-08T10:45:08,080 INFO [Time-limited test {}] mob.MobFileCache(128): MobFileCache disabled 2025-01-08T10:45:08,082 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T10:45:08,086 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T10:45:08,091 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=regionserver:45003 connecting to ZooKeeper ensemble=127.0.0.1:52367 2025-01-08T10:45:08,096 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:450030x0, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2025-01-08T10:45:08,098 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:450030x0, quorum=127.0.0.1:52367, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2025-01-08T10:45:08,104 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:450030x0, quorum=127.0.0.1:52367, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2025-01-08T10:45:08,107 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:45003-0x1000c1ee94b0001 connected 2025-01-08T10:45:08,109 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2025-01-08T10:45:08,121 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=45003 2025-01-08T10:45:08,122 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=45003 2025-01-08T10:45:08,134 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=45003 2025-01-08T10:45:08,140 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=45003 2025-01-08T10:45:08,141 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=45003 2025-01-08T10:45:08,167 INFO [Time-limited test {}] client.ConnectionUtils(129): regionserver/67f42ee69b4e:0 server-side Connection retries=45 2025-01-08T10:45:08,168 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2025-01-08T10:45:08,169 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2025-01-08T10:45:08,169 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2025-01-08T10:45:08,169 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2025-01-08T10:45:08,170 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2025-01-08T10:45:08,170 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2025-01-08T10:45:08,170 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2025-01-08T10:45:08,175 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:36283 2025-01-08T10:45:08,176 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2025-01-08T10:45:08,183 INFO [Time-limited test {}] mob.MobFileCache(128): MobFileCache disabled 2025-01-08T10:45:08,185 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T10:45:08,188 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T10:45:08,193 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=regionserver:36283 connecting to ZooKeeper ensemble=127.0.0.1:52367 2025-01-08T10:45:08,202 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:362830x0, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2025-01-08T10:45:08,202 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:362830x0, quorum=127.0.0.1:52367, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2025-01-08T10:45:08,207 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:362830x0, quorum=127.0.0.1:52367, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2025-01-08T10:45:08,209 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:362830x0, quorum=127.0.0.1:52367, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2025-01-08T10:45:08,211 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:36283-0x1000c1ee94b0002 connected 2025-01-08T10:45:08,221 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=36283 2025-01-08T10:45:08,222 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=36283 2025-01-08T10:45:08,226 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=36283 2025-01-08T10:45:08,230 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=36283 2025-01-08T10:45:08,230 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=36283 2025-01-08T10:45:08,260 INFO [Time-limited test {}] client.ConnectionUtils(129): regionserver/67f42ee69b4e:0 server-side Connection retries=45 2025-01-08T10:45:08,261 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2025-01-08T10:45:08,261 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2025-01-08T10:45:08,261 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2025-01-08T10:45:08,261 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2025-01-08T10:45:08,261 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2025-01-08T10:45:08,262 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2025-01-08T10:45:08,262 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2025-01-08T10:45:08,264 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:38323 2025-01-08T10:45:08,266 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2025-01-08T10:45:08,271 INFO [Time-limited test {}] mob.MobFileCache(128): MobFileCache disabled 2025-01-08T10:45:08,273 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T10:45:08,276 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T10:45:08,282 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=regionserver:38323 connecting to ZooKeeper ensemble=127.0.0.1:52367 2025-01-08T10:45:08,295 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:383230x0, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2025-01-08T10:45:08,296 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:383230x0, quorum=127.0.0.1:52367, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2025-01-08T10:45:08,299 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:383230x0, quorum=127.0.0.1:52367, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2025-01-08T10:45:08,302 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:383230x0, quorum=127.0.0.1:52367, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2025-01-08T10:45:08,307 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:38323-0x1000c1ee94b0003 connected 2025-01-08T10:45:08,319 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=38323 2025-01-08T10:45:08,329 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=38323 2025-01-08T10:45:08,350 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=38323 2025-01-08T10:45:08,370 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=38323 2025-01-08T10:45:08,377 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=38323 2025-01-08T10:45:08,397 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] master.HMaster(2460): Adding backup master ZNode /hbase/backup-masters/67f42ee69b4e,34551,1736333106957 2025-01-08T10:45:08,408 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2025-01-08T10:45:08,408 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2025-01-08T10:45:08,409 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2025-01-08T10:45:08,409 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2025-01-08T10:45:08,418 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/67f42ee69b4e,34551,1736333106957 2025-01-08T10:45:08,426 DEBUG [M:0;67f42ee69b4e:34551 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;67f42ee69b4e:34551 2025-01-08T10:45:08,470 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2025-01-08T10:45:08,470 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2025-01-08T10:45:08,470 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:08,470 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:08,470 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2025-01-08T10:45:08,470 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:08,471 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2025-01-08T10:45:08,471 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:08,479 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(111): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2025-01-08T10:45:08,482 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2025-01-08T10:45:08,487 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/67f42ee69b4e,34551,1736333106957 from backup master directory 2025-01-08T10:45:08,492 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2025-01-08T10:45:08,492 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2025-01-08T10:45:08,492 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/67f42ee69b4e,34551,1736333106957 2025-01-08T10:45:08,493 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2025-01-08T10:45:08,493 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2025-01-08T10:45:08,493 WARN [master/67f42ee69b4e:0:becomeActiveMaster {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2025-01-08T10:45:08,494 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=67f42ee69b4e,34551,1736333106957 2025-01-08T10:45:08,497 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating data MemStoreChunkPool with chunk size 2 MB, max count 396, initial count 0 2025-01-08T10:45:08,502 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating index MemStoreChunkPool with chunk size 204.80 KB, max count 440, initial count 0 2025-01-08T10:45:08,593 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741826_1002 (size=42) 2025-01-08T10:45:08,594 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741826_1002 (size=42) 2025-01-08T10:45:08,594 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741826_1002 (size=42) 2025-01-08T10:45:08,597 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] util.FSUtils(639): Created cluster ID file at hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/hbase.id with ID: 9fdc568b-f642-45bd-a97e-7c716b7dfe9a 2025-01-08T10:45:08,649 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T10:45:08,686 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:08,686 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:08,686 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:08,686 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:08,733 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741827_1003 (size=196) 2025-01-08T10:45:08,734 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741827_1003 (size=196) 2025-01-08T10:45:08,734 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741827_1003 (size=196) 2025-01-08T10:45:08,753 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] region.MasterRegion(372): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T10:45:08,755 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2025-01-08T10:45:08,780 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(244): No decryptEncryptedDataEncryptionKey method in DFSClient, should be hadoop version with HDFS-12396 java.lang.NoSuchMethodException: org.apache.hadoop.hdfs.DFSClient.decryptEncryptedDataEncryptionKey(org.apache.hadoop.fs.FileEncryptionInfo) at java.lang.Class.getDeclaredMethod(Class.java:2675) ~[?:?] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelperWithoutHDFS12396(FanOutOneBlockAsyncDFSOutputSaslHelper.java:183) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelper(FanOutOneBlockAsyncDFSOutputSaslHelper.java:242) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.(FanOutOneBlockAsyncDFSOutputSaslHelper.java:253) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at java.lang.Class.forName0(Native Method) ~[?:?] at java.lang.Class.forName(Class.java:375) ~[?:?] at org.apache.hadoop.hbase.wal.AsyncFSWALProvider.load(AsyncFSWALProvider.java:147) ~[classes/:?] at org.apache.hadoop.hbase.wal.WALFactory.getProviderClass(WALFactory.java:160) ~[classes/:?] at org.apache.hadoop.hbase.wal.WALFactory.getProvider(WALFactory.java:200) ~[classes/:?] at org.apache.hadoop.hbase.wal.WALFactory.(WALFactory.java:232) ~[classes/:?] at org.apache.hadoop.hbase.wal.WALFactory.(WALFactory.java:207) ~[classes/:?] at org.apache.hadoop.hbase.master.region.MasterRegion.create(MasterRegion.java:402) ~[classes/:?] at org.apache.hadoop.hbase.master.region.MasterRegionFactory.create(MasterRegionFactory.java:135) ~[classes/:?] at org.apache.hadoop.hbase.master.HMaster.finishActiveMasterInitialization(HMaster.java:984) ~[classes/:?] at org.apache.hadoop.hbase.master.HMaster.startActiveMasterManager(HMaster.java:2485) ~[classes/:?] at org.apache.hadoop.hbase.master.HMaster.lambda$run$0(HMaster.java:601) ~[classes/:?] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) ~[classes/:?] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) ~[classes/:?] at org.apache.hadoop.hbase.master.HMaster.lambda$run$1(HMaster.java:598) ~[classes/:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2025-01-08T10:45:08,787 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2025-01-08T10:45:08,881 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741828_1004 (size=1189) 2025-01-08T10:45:08,881 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741828_1004 (size=1189) 2025-01-08T10:45:08,882 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741828_1004 (size=1189) 2025-01-08T10:45:08,913 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] regionserver.HRegion(7124): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/MasterData/data/master/store 2025-01-08T10:45:08,953 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741829_1005 (size=34) 2025-01-08T10:45:08,953 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741829_1005 (size=34) 2025-01-08T10:45:08,954 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741829_1005 (size=34) 2025-01-08T10:45:08,960 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] throttle.StoreHotnessProtector(112): StoreHotnessProtector is disabled. Set hbase.region.store.parallel.put.limit > 0 to enable, which may help mitigate load under heavy write pressure. 2025-01-08T10:45:08,961 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:45:08,962 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2025-01-08T10:45:08,962 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2025-01-08T10:45:08,962 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2025-01-08T10:45:08,962 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2025-01-08T10:45:08,962 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2025-01-08T10:45:08,962 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] regionserver.HRegion(1922): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2025-01-08T10:45:08,963 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2025-01-08T10:45:08,966 WARN [master/67f42ee69b4e:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/MasterData/data/master/store/.initializing 2025-01-08T10:45:08,966 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/MasterData/WALs/67f42ee69b4e,34551,1736333106957 2025-01-08T10:45:08,975 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2025-01-08T10:45:08,990 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=67f42ee69b4e%2C34551%2C1736333106957, suffix=, logDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/MasterData/WALs/67f42ee69b4e,34551,1736333106957, archiveDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/MasterData/oldWALs, maxLogs=10 2025-01-08T10:45:09,018 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(617): When create output stream for /user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/MasterData/WALs/67f42ee69b4e,34551,1736333106957/67f42ee69b4e%2C34551%2C1736333106957.1736333108999, exclude list is [], retry=0 2025-01-08T10:45:09,037 DEBUG [RS-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:43487,DS-96de2396-f7dd-4dd7-805c-c34c70b41204,DISK] 2025-01-08T10:45:09,037 DEBUG [RS-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:44473,DS-687c546a-3440-42d4-b346-ee987087ce43,DISK] 2025-01-08T10:45:09,037 DEBUG [RS-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45285,DS-5709618f-16cd-4409-9a05-d0991502eaf3,DISK] 2025-01-08T10:45:09,040 DEBUG [RS-EventLoopGroup-5-1 {}] asyncfs.ProtobufDecoder(117): Hadoop 3.3 and above shades protobuf. 2025-01-08T10:45:09,079 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/MasterData/WALs/67f42ee69b4e,34551,1736333106957/67f42ee69b4e%2C34551%2C1736333106957.1736333108999 2025-01-08T10:45:09,080 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] wal.AbstractFSWAL(925): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:34377:34377),(127.0.0.1/127.0.0.1:41419:41419),(127.0.0.1/127.0.0.1:40217:40217)] 2025-01-08T10:45:09,081 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] regionserver.HRegion(7285): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2025-01-08T10:45:09,082 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:45:09,087 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] regionserver.HRegion(7327): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2025-01-08T10:45:09,088 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] regionserver.HRegion(7330): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2025-01-08T10:45:09,137 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2025-01-08T10:45:09,165 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2025-01-08T10:45:09,170 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:09,173 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2025-01-08T10:45:09,173 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2025-01-08T10:45:09,176 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2025-01-08T10:45:09,177 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:09,178 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:45:09,178 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2025-01-08T10:45:09,181 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2025-01-08T10:45:09,181 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:09,182 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:45:09,183 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2025-01-08T10:45:09,185 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2025-01-08T10:45:09,186 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:09,186 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:45:09,190 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2025-01-08T10:45:09,191 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2025-01-08T10:45:09,201 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2025-01-08T10:45:09,205 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] regionserver.HRegion(1085): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2025-01-08T10:45:09,209 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:45:09,210 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] regionserver.HRegion(1102): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=69125556, jitterRate=0.030051052570343018}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2025-01-08T10:45:09,215 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] regionserver.HRegion(1001): Region open journal for 1595e783b53d99cd5eef43b6debb2682: 2025-01-08T10:45:09,216 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2025-01-08T10:45:09,248 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2391c0fd, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:45:09,283 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] master.HMaster(893): No meta location available on zookeeper, skip migrating... 2025-01-08T10:45:09,295 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2025-01-08T10:45:09,296 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(633): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2025-01-08T10:45:09,298 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2025-01-08T10:45:09,299 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(653): Recovered RegionProcedureStore lease in 1 msec 2025-01-08T10:45:09,304 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(667): Loaded RegionProcedureStore in 4 msec 2025-01-08T10:45:09,305 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2025-01-08T10:45:09,331 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2025-01-08T10:45:09,343 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2025-01-08T10:45:09,345 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/balancer already deleted, retry=false 2025-01-08T10:45:09,347 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2025-01-08T10:45:09,349 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2025-01-08T10:45:09,350 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/normalizer already deleted, retry=false 2025-01-08T10:45:09,352 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2025-01-08T10:45:09,355 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2025-01-08T10:45:09,357 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/split already deleted, retry=false 2025-01-08T10:45:09,358 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2025-01-08T10:45:09,359 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/merge already deleted, retry=false 2025-01-08T10:45:09,369 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2025-01-08T10:45:09,370 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/snapshot-cleanup already deleted, retry=false 2025-01-08T10:45:09,373 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2025-01-08T10:45:09,373 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2025-01-08T10:45:09,373 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2025-01-08T10:45:09,373 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2025-01-08T10:45:09,374 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:09,374 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:09,374 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:09,374 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:09,375 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] master.HMaster(837): Active/primary master=67f42ee69b4e,34551,1736333106957, sessionid=0x1000c1ee94b0000, setting cluster-up flag (Was=false) 2025-01-08T10:45:09,387 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:09,387 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:09,387 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:09,387 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:09,392 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2025-01-08T10:45:09,394 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=67f42ee69b4e,34551,1736333106957 2025-01-08T10:45:09,399 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:09,399 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:09,399 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:09,399 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:09,405 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2025-01-08T10:45:09,406 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=67f42ee69b4e,34551,1736333106957 2025-01-08T10:45:09,505 DEBUG [RS:0;67f42ee69b4e:45003 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;67f42ee69b4e:45003 2025-01-08T10:45:09,505 DEBUG [RS:1;67f42ee69b4e:36283 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;67f42ee69b4e:36283 2025-01-08T10:45:09,507 INFO [RS:0;67f42ee69b4e:45003 {}] regionserver.HRegionServer(1008): ClusterId : 9fdc568b-f642-45bd-a97e-7c716b7dfe9a 2025-01-08T10:45:09,507 INFO [RS:1;67f42ee69b4e:36283 {}] regionserver.HRegionServer(1008): ClusterId : 9fdc568b-f642-45bd-a97e-7c716b7dfe9a 2025-01-08T10:45:09,510 DEBUG [RS:1;67f42ee69b4e:36283 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2025-01-08T10:45:09,510 DEBUG [RS:0;67f42ee69b4e:45003 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2025-01-08T10:45:09,511 DEBUG [RS:2;67f42ee69b4e:38323 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;67f42ee69b4e:38323 2025-01-08T10:45:09,514 INFO [RS:2;67f42ee69b4e:38323 {}] regionserver.HRegionServer(1008): ClusterId : 9fdc568b-f642-45bd-a97e-7c716b7dfe9a 2025-01-08T10:45:09,514 DEBUG [RS:2;67f42ee69b4e:38323 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2025-01-08T10:45:09,516 DEBUG [RS:0;67f42ee69b4e:45003 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2025-01-08T10:45:09,516 DEBUG [RS:1;67f42ee69b4e:36283 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2025-01-08T10:45:09,516 DEBUG [RS:0;67f42ee69b4e:45003 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2025-01-08T10:45:09,516 DEBUG [RS:1;67f42ee69b4e:36283 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2025-01-08T10:45:09,519 DEBUG [RS:2;67f42ee69b4e:38323 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2025-01-08T10:45:09,519 DEBUG [RS:2;67f42ee69b4e:38323 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2025-01-08T10:45:09,519 DEBUG [RS:0;67f42ee69b4e:45003 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2025-01-08T10:45:09,520 DEBUG [RS:1;67f42ee69b4e:36283 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2025-01-08T10:45:09,520 DEBUG [RS:1;67f42ee69b4e:36283 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@64e11c1b, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:45:09,520 DEBUG [RS:0;67f42ee69b4e:45003 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@41f57ebb, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:45:09,522 DEBUG [RS:1;67f42ee69b4e:36283 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1b3055cf, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=67f42ee69b4e/172.17.0.2:0 2025-01-08T10:45:09,522 DEBUG [RS:0;67f42ee69b4e:45003 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1a017334, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=67f42ee69b4e/172.17.0.2:0 2025-01-08T10:45:09,523 DEBUG [RS:2;67f42ee69b4e:38323 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2025-01-08T10:45:09,523 DEBUG [RS:2;67f42ee69b4e:38323 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@21a9e220, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:45:09,526 DEBUG [RS:2;67f42ee69b4e:38323 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@3ac4f474, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=67f42ee69b4e/172.17.0.2:0 2025-01-08T10:45:09,528 INFO [RS:2;67f42ee69b4e:38323 {}] regionserver.RegionServerCoprocessorHost(67): System coprocessor loading is enabled 2025-01-08T10:45:09,528 INFO [RS:0;67f42ee69b4e:45003 {}] regionserver.RegionServerCoprocessorHost(67): System coprocessor loading is enabled 2025-01-08T10:45:09,528 INFO [RS:1;67f42ee69b4e:36283 {}] regionserver.RegionServerCoprocessorHost(67): System coprocessor loading is enabled 2025-01-08T10:45:09,528 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] master.HMaster(3405): Registered master coprocessor service: service=AccessControlService 2025-01-08T10:45:09,528 INFO [RS:2;67f42ee69b4e:38323 {}] regionserver.RegionServerCoprocessorHost(68): Table coprocessor loading is enabled 2025-01-08T10:45:09,528 INFO [RS:1;67f42ee69b4e:36283 {}] regionserver.RegionServerCoprocessorHost(68): Table coprocessor loading is enabled 2025-01-08T10:45:09,528 INFO [RS:0;67f42ee69b4e:45003 {}] regionserver.RegionServerCoprocessorHost(68): Table coprocessor loading is enabled 2025-01-08T10:45:09,532 DEBUG [RS:2;67f42ee69b4e:38323 {}] regionserver.HRegionServer(900): Registered regionserver coprocessor executorService: executorService=AccessControlService 2025-01-08T10:45:09,532 DEBUG [RS:1;67f42ee69b4e:36283 {}] regionserver.HRegionServer(900): Registered regionserver coprocessor executorService: executorService=AccessControlService 2025-01-08T10:45:09,532 DEBUG [RS:0;67f42ee69b4e:45003 {}] regionserver.HRegionServer(900): Registered regionserver coprocessor executorService: executorService=AccessControlService 2025-01-08T10:45:09,533 INFO [RS:0;67f42ee69b4e:45003 {}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:45:09,533 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:45:09,533 INFO [RS:2;67f42ee69b4e:38323 {}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:45:09,533 INFO [RS:1;67f42ee69b4e:36283 {}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:45:09,533 DEBUG [RS:0;67f42ee69b4e:45003 {}] regionserver.HRegionServer(1090): About to register with Master. 2025-01-08T10:45:09,533 DEBUG [RS:2;67f42ee69b4e:38323 {}] regionserver.HRegionServer(1090): About to register with Master. 2025-01-08T10:45:09,533 DEBUG [RS:1;67f42ee69b4e:36283 {}] regionserver.HRegionServer(1090): About to register with Master. 2025-01-08T10:45:09,534 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.SecureTestUtil$MasterSyncObserver loaded, priority=536870912. 2025-01-08T10:45:09,536 INFO [RS:1;67f42ee69b4e:36283 {}] regionserver.HRegionServer(3073): reportForDuty to master=67f42ee69b4e,34551,1736333106957 with isa=67f42ee69b4e/172.17.0.2:36283, startcode=1736333108166 2025-01-08T10:45:09,539 INFO [RS:0;67f42ee69b4e:45003 {}] regionserver.HRegionServer(3073): reportForDuty to master=67f42ee69b4e,34551,1736333106957 with isa=67f42ee69b4e/172.17.0.2:45003, startcode=1736333108039 2025-01-08T10:45:09,539 INFO [RS:2;67f42ee69b4e:38323 {}] regionserver.HRegionServer(3073): reportForDuty to master=67f42ee69b4e,34551,1736333106957 with isa=67f42ee69b4e/172.17.0.2:38323, startcode=1736333108259 2025-01-08T10:45:09,551 DEBUG [RS:0;67f42ee69b4e:45003 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2025-01-08T10:45:09,551 DEBUG [RS:2;67f42ee69b4e:38323 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2025-01-08T10:45:09,551 DEBUG [RS:1;67f42ee69b4e:36283 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2025-01-08T10:45:09,593 INFO [RS-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:48835, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=RegionServerStatusService 2025-01-08T10:45:09,593 INFO [RS-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:51095, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.1 (auth:SIMPLE), service=RegionServerStatusService 2025-01-08T10:45:09,593 INFO [RS-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:47637, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.2 (auth:SIMPLE), service=RegionServerStatusService 2025-01-08T10:45:09,597 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT; InitMetaProcedure table=hbase:meta 2025-01-08T10:45:09,600 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=34551 {}] ipc.MetricsHBaseServer(152): Unknown exception type org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server is not running yet at org.apache.hadoop.hbase.master.HMaster.checkServiceStarted(HMaster.java:3295) ~[classes/:?] at org.apache.hadoop.hbase.master.MasterRpcServices.regionServerStartup(MasterRpcServices.java:593) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$2.callBlockingMethod(RegionServerStatusProtos.java:16714) ~[hbase-protocol-shaded-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:443) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:105) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) ~[classes/:2.7.0-SNAPSHOT] 2025-01-08T10:45:09,604 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(575): slop=0.2 2025-01-08T10:45:09,606 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=34551 {}] ipc.MetricsHBaseServer(152): Unknown exception type org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server is not running yet at org.apache.hadoop.hbase.master.HMaster.checkServiceStarted(HMaster.java:3295) ~[classes/:?] at org.apache.hadoop.hbase.master.MasterRpcServices.regionServerStartup(MasterRpcServices.java:593) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$2.callBlockingMethod(RegionServerStatusProtos.java:16714) ~[hbase-protocol-shaded-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:443) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:105) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) ~[classes/:2.7.0-SNAPSHOT] 2025-01-08T10:45:09,606 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=34551 {}] ipc.MetricsHBaseServer(152): Unknown exception type org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server is not running yet at org.apache.hadoop.hbase.master.HMaster.checkServiceStarted(HMaster.java:3295) ~[classes/:?] at org.apache.hadoop.hbase.master.MasterRpcServices.regionServerStartup(MasterRpcServices.java:593) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$2.callBlockingMethod(RegionServerStatusProtos.java:16714) ~[hbase-protocol-shaded-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:443) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:105) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) ~[classes/:2.7.0-SNAPSHOT] 2025-01-08T10:45:09,609 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(294): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2025-01-08T10:45:09,616 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] balancer.RegionLocationFinder(146): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 67f42ee69b4e,34551,1736333106957 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2025-01-08T10:45:09,620 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/67f42ee69b4e:0, corePoolSize=5, maxPoolSize=5 2025-01-08T10:45:09,620 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/67f42ee69b4e:0, corePoolSize=5, maxPoolSize=5 2025-01-08T10:45:09,621 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/67f42ee69b4e:0, corePoolSize=5, maxPoolSize=5 2025-01-08T10:45:09,621 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/67f42ee69b4e:0, corePoolSize=5, maxPoolSize=5 2025-01-08T10:45:09,621 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/67f42ee69b4e:0, corePoolSize=10, maxPoolSize=10 2025-01-08T10:45:09,621 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,621 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/67f42ee69b4e:0, corePoolSize=2, maxPoolSize=2 2025-01-08T10:45:09,621 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,631 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, locked=true; InitMetaProcedure table=hbase:meta 2025-01-08T10:45:09,631 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(75): BOOTSTRAP: creating hbase:meta region 2025-01-08T10:45:09,631 DEBUG [RS:0;67f42ee69b4e:45003 {}] regionserver.HRegionServer(3097): Master is not running yet 2025-01-08T10:45:09,631 DEBUG [RS:1;67f42ee69b4e:36283 {}] regionserver.HRegionServer(3097): Master is not running yet 2025-01-08T10:45:09,632 WARN [RS:0;67f42ee69b4e:45003 {}] regionserver.HRegionServer(1099): reportForDuty failed; sleeping 100 ms and then retrying. 2025-01-08T10:45:09,632 WARN [RS:1;67f42ee69b4e:36283 {}] regionserver.HRegionServer(1099): reportForDuty failed; sleeping 100 ms and then retrying. 2025-01-08T10:45:09,631 DEBUG [RS:2;67f42ee69b4e:38323 {}] regionserver.HRegionServer(3097): Master is not running yet 2025-01-08T10:45:09,632 WARN [RS:2;67f42ee69b4e:38323 {}] regionserver.HRegionServer(1099): reportForDuty failed; sleeping 100 ms and then retrying. 2025-01-08T10:45:09,633 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1736333139633 2025-01-08T10:45:09,636 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2025-01-08T10:45:09,637 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:09,637 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2025-01-08T10:45:09,637 INFO [PEWorker-1 {}] util.FSTableDescriptors(133): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2025-01-08T10:45:09,641 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2025-01-08T10:45:09,642 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2025-01-08T10:45:09,642 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2025-01-08T10:45:09,642 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2025-01-08T10:45:09,644 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:09,649 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2025-01-08T10:45:09,651 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2025-01-08T10:45:09,652 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2025-01-08T10:45:09,659 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2025-01-08T10:45:09,660 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2025-01-08T10:45:09,663 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741831_1007 (size=1039) 2025-01-08T10:45:09,664 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741831_1007 (size=1039) 2025-01-08T10:45:09,664 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741831_1007 (size=1039) 2025-01-08T10:45:09,666 INFO [PEWorker-1 {}] util.FSTableDescriptors(140): Updated hbase:meta table descriptor to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1039 2025-01-08T10:45:09,666 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/67f42ee69b4e:0:becomeActiveMaster-HFileCleaner.large.0-1736333109662,5,FailOnTimeoutGroup] 2025-01-08T10:45:09,667 INFO [PEWorker-1 {}] regionserver.HRegion(7106): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:45:09,674 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/67f42ee69b4e:0:becomeActiveMaster-HFileCleaner.small.0-1736333109667,5,FailOnTimeoutGroup] 2025-01-08T10:45:09,675 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:09,675 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] master.HMaster(1691): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2025-01-08T10:45:09,677 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:09,677 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:09,695 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741832_1008 (size=32) 2025-01-08T10:45:09,695 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741832_1008 (size=32) 2025-01-08T10:45:09,696 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741832_1008 (size=32) 2025-01-08T10:45:09,698 DEBUG [PEWorker-1 {}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:45:09,701 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2025-01-08T10:45:09,704 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2025-01-08T10:45:09,704 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:09,705 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2025-01-08T10:45:09,706 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2025-01-08T10:45:09,708 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2025-01-08T10:45:09,708 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:09,709 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2025-01-08T10:45:09,709 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2025-01-08T10:45:09,712 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2025-01-08T10:45:09,712 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:09,713 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2025-01-08T10:45:09,714 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/meta/1588230740 2025-01-08T10:45:09,715 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/meta/1588230740 2025-01-08T10:45:09,718 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (42.7 M)) instead. 2025-01-08T10:45:09,721 DEBUG [PEWorker-1 {}] regionserver.HRegion(1085): writing seq id for 1588230740 2025-01-08T10:45:09,725 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:45:09,726 INFO [PEWorker-1 {}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=66588663, jitterRate=-0.007751598954200745}}}, FlushLargeStoresPolicy{flushSizeLowerBound=44739242} 2025-01-08T10:45:09,728 DEBUG [PEWorker-1 {}] regionserver.HRegion(1001): Region open journal for 1588230740: 2025-01-08T10:45:09,728 DEBUG [PEWorker-1 {}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2025-01-08T10:45:09,728 INFO [PEWorker-1 {}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2025-01-08T10:45:09,728 DEBUG [PEWorker-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2025-01-08T10:45:09,728 DEBUG [PEWorker-1 {}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2025-01-08T10:45:09,729 DEBUG [PEWorker-1 {}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2025-01-08T10:45:09,730 INFO [PEWorker-1 {}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2025-01-08T10:45:09,730 DEBUG [PEWorker-1 {}] regionserver.HRegion(1635): Region close journal for 1588230740: 2025-01-08T10:45:09,733 INFO [RS:1;67f42ee69b4e:36283 {}] regionserver.HRegionServer(3073): reportForDuty to master=67f42ee69b4e,34551,1736333106957 with isa=67f42ee69b4e/172.17.0.2:36283, startcode=1736333108166 2025-01-08T10:45:09,734 INFO [RS:2;67f42ee69b4e:38323 {}] regionserver.HRegionServer(3073): reportForDuty to master=67f42ee69b4e,34551,1736333106957 with isa=67f42ee69b4e/172.17.0.2:38323, startcode=1736333108259 2025-01-08T10:45:09,734 INFO [RS:0;67f42ee69b4e:45003 {}] regionserver.HRegionServer(3073): reportForDuty to master=67f42ee69b4e,34551,1736333106957 with isa=67f42ee69b4e/172.17.0.2:45003, startcode=1736333108039 2025-01-08T10:45:09,734 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, locked=true; InitMetaProcedure table=hbase:meta 2025-01-08T10:45:09,734 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(107): Going to assign meta 2025-01-08T10:45:09,735 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=34551 {}] master.ServerManager(332): Checking decommissioned status of RegionServer 67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:09,739 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=34551 {}] master.ServerManager(486): Registering regionserver=67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:09,741 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2025-01-08T10:45:09,748 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=34551 {}] master.ServerManager(332): Checking decommissioned status of RegionServer 67f42ee69b4e,38323,1736333108259 2025-01-08T10:45:09,748 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=34551 {}] master.ServerManager(486): Registering regionserver=67f42ee69b4e,38323,1736333108259 2025-01-08T10:45:09,748 DEBUG [RS:1;67f42ee69b4e:36283 {}] regionserver.HRegionServer(1725): Config from master: hbase.rootdir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:45:09,748 DEBUG [RS:1;67f42ee69b4e:36283 {}] regionserver.HRegionServer(1725): Config from master: fs.defaultFS=hdfs://localhost:37495 2025-01-08T10:45:09,748 DEBUG [RS:1;67f42ee69b4e:36283 {}] regionserver.HRegionServer(1725): Config from master: hbase.master.info.port=-1 2025-01-08T10:45:09,751 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2025-01-08T10:45:09,751 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=34551 {}] master.ServerManager(332): Checking decommissioned status of RegionServer 67f42ee69b4e,45003,1736333108039 2025-01-08T10:45:09,751 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=34551 {}] master.ServerManager(486): Registering regionserver=67f42ee69b4e,45003,1736333108039 2025-01-08T10:45:09,752 DEBUG [RS:2;67f42ee69b4e:38323 {}] regionserver.HRegionServer(1725): Config from master: hbase.rootdir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:45:09,752 DEBUG [RS:2;67f42ee69b4e:38323 {}] regionserver.HRegionServer(1725): Config from master: fs.defaultFS=hdfs://localhost:37495 2025-01-08T10:45:09,752 DEBUG [RS:2;67f42ee69b4e:38323 {}] regionserver.HRegionServer(1725): Config from master: hbase.master.info.port=-1 2025-01-08T10:45:09,753 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2025-01-08T10:45:09,754 INFO [PEWorker-3 {}] assignment.TransitRegionStateProcedure(264): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2025-01-08T10:45:09,755 DEBUG [RS:0;67f42ee69b4e:45003 {}] regionserver.HRegionServer(1725): Config from master: hbase.rootdir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:45:09,755 DEBUG [RS:0;67f42ee69b4e:45003 {}] regionserver.HRegionServer(1725): Config from master: fs.defaultFS=hdfs://localhost:37495 2025-01-08T10:45:09,755 DEBUG [RS:0;67f42ee69b4e:45003 {}] regionserver.HRegionServer(1725): Config from master: hbase.master.info.port=-1 2025-01-08T10:45:09,758 DEBUG [RS:1;67f42ee69b4e:36283 {}] zookeeper.ZKUtil(111): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:09,758 WARN [RS:1;67f42ee69b4e:36283 {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2025-01-08T10:45:09,758 INFO [RS:1;67f42ee69b4e:36283 {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2025-01-08T10:45:09,758 DEBUG [RS:1;67f42ee69b4e:36283 {}] regionserver.HRegionServer(2100): logDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/WALs/67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:09,759 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2025-01-08T10:45:09,759 DEBUG [RS:2;67f42ee69b4e:38323 {}] zookeeper.ZKUtil(111): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/67f42ee69b4e,38323,1736333108259 2025-01-08T10:45:09,759 WARN [RS:2;67f42ee69b4e:38323 {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2025-01-08T10:45:09,759 DEBUG [RS:0;67f42ee69b4e:45003 {}] zookeeper.ZKUtil(111): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/67f42ee69b4e,45003,1736333108039 2025-01-08T10:45:09,760 WARN [RS:0;67f42ee69b4e:45003 {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2025-01-08T10:45:09,760 INFO [RS:2;67f42ee69b4e:38323 {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2025-01-08T10:45:09,760 INFO [RS:0;67f42ee69b4e:45003 {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2025-01-08T10:45:09,760 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [67f42ee69b4e,36283,1736333108166] 2025-01-08T10:45:09,760 DEBUG [RS:2;67f42ee69b4e:38323 {}] regionserver.HRegionServer(2100): logDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/WALs/67f42ee69b4e,38323,1736333108259 2025-01-08T10:45:09,760 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [67f42ee69b4e,38323,1736333108259] 2025-01-08T10:45:09,760 DEBUG [RS:0;67f42ee69b4e:45003 {}] regionserver.HRegionServer(2100): logDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/WALs/67f42ee69b4e,45003,1736333108039 2025-01-08T10:45:09,761 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [67f42ee69b4e,45003,1736333108039] 2025-01-08T10:45:09,775 DEBUG [RS:2;67f42ee69b4e:38323 {}] regionserver.Replication(140): Replication stats-in-log period=300 seconds 2025-01-08T10:45:09,775 DEBUG [RS:1;67f42ee69b4e:36283 {}] regionserver.Replication(140): Replication stats-in-log period=300 seconds 2025-01-08T10:45:09,775 DEBUG [RS:0;67f42ee69b4e:45003 {}] regionserver.Replication(140): Replication stats-in-log period=300 seconds 2025-01-08T10:45:09,788 INFO [RS:1;67f42ee69b4e:36283 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2025-01-08T10:45:09,788 INFO [RS:2;67f42ee69b4e:38323 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2025-01-08T10:45:09,788 INFO [RS:0;67f42ee69b4e:45003 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2025-01-08T10:45:09,802 INFO [RS:2;67f42ee69b4e:38323 {}] regionserver.MemStoreFlusher(130): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2025-01-08T10:45:09,803 INFO [RS:1;67f42ee69b4e:36283 {}] regionserver.MemStoreFlusher(130): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2025-01-08T10:45:09,803 INFO [RS:0;67f42ee69b4e:45003 {}] regionserver.MemStoreFlusher(130): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2025-01-08T10:45:09,806 INFO [RS:0;67f42ee69b4e:45003 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2025-01-08T10:45:09,806 INFO [RS:2;67f42ee69b4e:38323 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2025-01-08T10:45:09,807 INFO [RS:0;67f42ee69b4e:45003 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:09,807 INFO [RS:1;67f42ee69b4e:36283 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2025-01-08T10:45:09,807 INFO [RS:2;67f42ee69b4e:38323 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:09,807 INFO [RS:1;67f42ee69b4e:36283 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:09,807 INFO [RS:0;67f42ee69b4e:45003 {}] regionserver.HRegionServer$CompactionChecker(1988): CompactionChecker runs every PT1S 2025-01-08T10:45:09,807 INFO [RS:2;67f42ee69b4e:38323 {}] regionserver.HRegionServer$CompactionChecker(1988): CompactionChecker runs every PT1S 2025-01-08T10:45:09,807 INFO [RS:1;67f42ee69b4e:36283 {}] regionserver.HRegionServer$CompactionChecker(1988): CompactionChecker runs every PT1S 2025-01-08T10:45:09,815 INFO [RS:0;67f42ee69b4e:45003 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:09,815 INFO [RS:2;67f42ee69b4e:38323 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:09,815 INFO [RS:1;67f42ee69b4e:36283 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:09,815 DEBUG [RS:1;67f42ee69b4e:36283 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,815 DEBUG [RS:2;67f42ee69b4e:38323 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,815 DEBUG [RS:0;67f42ee69b4e:45003 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,816 DEBUG [RS:1;67f42ee69b4e:36283 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,816 DEBUG [RS:2;67f42ee69b4e:38323 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,816 DEBUG [RS:1;67f42ee69b4e:36283 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,816 DEBUG [RS:2;67f42ee69b4e:38323 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,816 DEBUG [RS:1;67f42ee69b4e:36283 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,816 DEBUG [RS:0;67f42ee69b4e:45003 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,816 DEBUG [RS:2;67f42ee69b4e:38323 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,816 DEBUG [RS:0;67f42ee69b4e:45003 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,816 DEBUG [RS:1;67f42ee69b4e:36283 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,816 DEBUG [RS:2;67f42ee69b4e:38323 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,816 DEBUG [RS:0;67f42ee69b4e:45003 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,816 DEBUG [RS:1;67f42ee69b4e:36283 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/67f42ee69b4e:0, corePoolSize=2, maxPoolSize=2 2025-01-08T10:45:09,816 DEBUG [RS:2;67f42ee69b4e:38323 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/67f42ee69b4e:0, corePoolSize=2, maxPoolSize=2 2025-01-08T10:45:09,816 DEBUG [RS:1;67f42ee69b4e:36283 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,816 DEBUG [RS:0;67f42ee69b4e:45003 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,816 DEBUG [RS:1;67f42ee69b4e:36283 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,816 DEBUG [RS:2;67f42ee69b4e:38323 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,816 DEBUG [RS:0;67f42ee69b4e:45003 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/67f42ee69b4e:0, corePoolSize=2, maxPoolSize=2 2025-01-08T10:45:09,817 DEBUG [RS:1;67f42ee69b4e:36283 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,817 DEBUG [RS:0;67f42ee69b4e:45003 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,817 DEBUG [RS:2;67f42ee69b4e:38323 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,817 DEBUG [RS:1;67f42ee69b4e:36283 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,817 DEBUG [RS:0;67f42ee69b4e:45003 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,817 DEBUG [RS:1;67f42ee69b4e:36283 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,817 DEBUG [RS:2;67f42ee69b4e:38323 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,817 DEBUG [RS:0;67f42ee69b4e:45003 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,817 DEBUG [RS:1;67f42ee69b4e:36283 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0, corePoolSize=3, maxPoolSize=3 2025-01-08T10:45:09,817 DEBUG [RS:0;67f42ee69b4e:45003 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,817 DEBUG [RS:2;67f42ee69b4e:38323 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,817 DEBUG [RS:1;67f42ee69b4e:36283 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/67f42ee69b4e:0, corePoolSize=3, maxPoolSize=3 2025-01-08T10:45:09,817 DEBUG [RS:0;67f42ee69b4e:45003 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,817 DEBUG [RS:2;67f42ee69b4e:38323 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/67f42ee69b4e:0, corePoolSize=1, maxPoolSize=1 2025-01-08T10:45:09,817 DEBUG [RS:0;67f42ee69b4e:45003 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0, corePoolSize=3, maxPoolSize=3 2025-01-08T10:45:09,817 DEBUG [RS:2;67f42ee69b4e:38323 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0, corePoolSize=3, maxPoolSize=3 2025-01-08T10:45:09,817 DEBUG [RS:0;67f42ee69b4e:45003 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/67f42ee69b4e:0, corePoolSize=3, maxPoolSize=3 2025-01-08T10:45:09,817 DEBUG [RS:2;67f42ee69b4e:38323 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/67f42ee69b4e:0, corePoolSize=3, maxPoolSize=3 2025-01-08T10:45:09,819 INFO [RS:1;67f42ee69b4e:36283 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:09,819 INFO [RS:1;67f42ee69b4e:36283 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:09,819 INFO [RS:1;67f42ee69b4e:36283 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:09,819 INFO [RS:1;67f42ee69b4e:36283 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:09,819 INFO [RS:1;67f42ee69b4e:36283 {}] hbase.ChoreService(168): Chore ScheduledChore name=67f42ee69b4e,36283,1736333108166-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2025-01-08T10:45:09,819 INFO [RS:0;67f42ee69b4e:45003 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:09,819 INFO [RS:2;67f42ee69b4e:38323 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:09,820 INFO [RS:0;67f42ee69b4e:45003 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:09,820 INFO [RS:2;67f42ee69b4e:38323 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:09,820 INFO [RS:0;67f42ee69b4e:45003 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:09,820 INFO [RS:0;67f42ee69b4e:45003 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:09,820 INFO [RS:2;67f42ee69b4e:38323 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:09,820 INFO [RS:0;67f42ee69b4e:45003 {}] hbase.ChoreService(168): Chore ScheduledChore name=67f42ee69b4e,45003,1736333108039-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2025-01-08T10:45:09,820 INFO [RS:2;67f42ee69b4e:38323 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:09,820 INFO [RS:2;67f42ee69b4e:38323 {}] hbase.ChoreService(168): Chore ScheduledChore name=67f42ee69b4e,38323,1736333108259-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2025-01-08T10:45:09,841 INFO [RS:1;67f42ee69b4e:36283 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2025-01-08T10:45:09,843 INFO [RS:1;67f42ee69b4e:36283 {}] hbase.ChoreService(168): Chore ScheduledChore name=67f42ee69b4e,36283,1736333108166-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:09,845 INFO [RS:0;67f42ee69b4e:45003 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2025-01-08T10:45:09,845 INFO [RS:2;67f42ee69b4e:38323 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2025-01-08T10:45:09,845 INFO [RS:2;67f42ee69b4e:38323 {}] hbase.ChoreService(168): Chore ScheduledChore name=67f42ee69b4e,38323,1736333108259-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:09,845 INFO [RS:0;67f42ee69b4e:45003 {}] hbase.ChoreService(168): Chore ScheduledChore name=67f42ee69b4e,45003,1736333108039-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:09,862 INFO [RS:1;67f42ee69b4e:36283 {}] regionserver.Replication(204): 67f42ee69b4e,36283,1736333108166 started 2025-01-08T10:45:09,862 INFO [RS:2;67f42ee69b4e:38323 {}] regionserver.Replication(204): 67f42ee69b4e,38323,1736333108259 started 2025-01-08T10:45:09,863 INFO [RS:1;67f42ee69b4e:36283 {}] regionserver.HRegionServer(1767): Serving as 67f42ee69b4e,36283,1736333108166, RpcServer on 67f42ee69b4e/172.17.0.2:36283, sessionid=0x1000c1ee94b0002 2025-01-08T10:45:09,863 INFO [RS:0;67f42ee69b4e:45003 {}] regionserver.Replication(204): 67f42ee69b4e,45003,1736333108039 started 2025-01-08T10:45:09,863 INFO [RS:2;67f42ee69b4e:38323 {}] regionserver.HRegionServer(1767): Serving as 67f42ee69b4e,38323,1736333108259, RpcServer on 67f42ee69b4e/172.17.0.2:38323, sessionid=0x1000c1ee94b0003 2025-01-08T10:45:09,863 INFO [RS:0;67f42ee69b4e:45003 {}] regionserver.HRegionServer(1767): Serving as 67f42ee69b4e,45003,1736333108039, RpcServer on 67f42ee69b4e/172.17.0.2:45003, sessionid=0x1000c1ee94b0001 2025-01-08T10:45:09,863 DEBUG [RS:2;67f42ee69b4e:38323 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2025-01-08T10:45:09,863 DEBUG [RS:0;67f42ee69b4e:45003 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2025-01-08T10:45:09,863 DEBUG [RS:1;67f42ee69b4e:36283 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2025-01-08T10:45:09,863 DEBUG [RS:2;67f42ee69b4e:38323 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 67f42ee69b4e,38323,1736333108259 2025-01-08T10:45:09,863 DEBUG [RS:0;67f42ee69b4e:45003 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 67f42ee69b4e,45003,1736333108039 2025-01-08T10:45:09,863 DEBUG [RS:1;67f42ee69b4e:36283 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:09,863 DEBUG [RS:0;67f42ee69b4e:45003 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '67f42ee69b4e,45003,1736333108039' 2025-01-08T10:45:09,863 DEBUG [RS:1;67f42ee69b4e:36283 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '67f42ee69b4e,36283,1736333108166' 2025-01-08T10:45:09,863 DEBUG [RS:2;67f42ee69b4e:38323 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '67f42ee69b4e,38323,1736333108259' 2025-01-08T10:45:09,863 DEBUG [RS:0;67f42ee69b4e:45003 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2025-01-08T10:45:09,863 DEBUG [RS:1;67f42ee69b4e:36283 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2025-01-08T10:45:09,863 DEBUG [RS:2;67f42ee69b4e:38323 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2025-01-08T10:45:09,864 DEBUG [RS:0;67f42ee69b4e:45003 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2025-01-08T10:45:09,864 DEBUG [RS:2;67f42ee69b4e:38323 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2025-01-08T10:45:09,864 DEBUG [RS:1;67f42ee69b4e:36283 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2025-01-08T10:45:09,865 DEBUG [RS:0;67f42ee69b4e:45003 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2025-01-08T10:45:09,865 DEBUG [RS:2;67f42ee69b4e:38323 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2025-01-08T10:45:09,865 DEBUG [RS:1;67f42ee69b4e:36283 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2025-01-08T10:45:09,865 DEBUG [RS:2;67f42ee69b4e:38323 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2025-01-08T10:45:09,865 DEBUG [RS:0;67f42ee69b4e:45003 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2025-01-08T10:45:09,865 DEBUG [RS:1;67f42ee69b4e:36283 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2025-01-08T10:45:09,865 DEBUG [RS:1;67f42ee69b4e:36283 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:09,865 DEBUG [RS:0;67f42ee69b4e:45003 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 67f42ee69b4e,45003,1736333108039 2025-01-08T10:45:09,865 DEBUG [RS:2;67f42ee69b4e:38323 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 67f42ee69b4e,38323,1736333108259 2025-01-08T10:45:09,865 DEBUG [RS:1;67f42ee69b4e:36283 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '67f42ee69b4e,36283,1736333108166' 2025-01-08T10:45:09,865 DEBUG [RS:1;67f42ee69b4e:36283 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2025-01-08T10:45:09,865 DEBUG [RS:2;67f42ee69b4e:38323 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '67f42ee69b4e,38323,1736333108259' 2025-01-08T10:45:09,865 DEBUG [RS:0;67f42ee69b4e:45003 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '67f42ee69b4e,45003,1736333108039' 2025-01-08T10:45:09,865 DEBUG [RS:0;67f42ee69b4e:45003 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2025-01-08T10:45:09,865 DEBUG [RS:2;67f42ee69b4e:38323 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2025-01-08T10:45:09,866 DEBUG [RS:1;67f42ee69b4e:36283 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2025-01-08T10:45:09,866 DEBUG [RS:2;67f42ee69b4e:38323 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2025-01-08T10:45:09,866 DEBUG [RS:0;67f42ee69b4e:45003 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2025-01-08T10:45:09,866 DEBUG [RS:1;67f42ee69b4e:36283 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2025-01-08T10:45:09,866 INFO [RS:1;67f42ee69b4e:36283 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2025-01-08T10:45:09,866 INFO [RS:1;67f42ee69b4e:36283 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2025-01-08T10:45:09,866 DEBUG [RS:0;67f42ee69b4e:45003 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2025-01-08T10:45:09,866 DEBUG [RS:2;67f42ee69b4e:38323 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2025-01-08T10:45:09,867 INFO [RS:0;67f42ee69b4e:45003 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2025-01-08T10:45:09,867 INFO [RS:2;67f42ee69b4e:38323 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2025-01-08T10:45:09,867 INFO [RS:0;67f42ee69b4e:45003 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2025-01-08T10:45:09,867 INFO [RS:2;67f42ee69b4e:38323 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2025-01-08T10:45:09,905 WARN [67f42ee69b4e:34551 {}] assignment.AssignmentManager(2423): No servers available; cannot place 1 unassigned regions. 2025-01-08T10:45:09,973 INFO [RS:1;67f42ee69b4e:36283 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2025-01-08T10:45:09,973 INFO [RS:2;67f42ee69b4e:38323 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2025-01-08T10:45:09,974 INFO [RS:0;67f42ee69b4e:45003 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2025-01-08T10:45:09,978 INFO [RS:1;67f42ee69b4e:36283 {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=67f42ee69b4e%2C36283%2C1736333108166, suffix=, logDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/WALs/67f42ee69b4e,36283,1736333108166, archiveDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/oldWALs, maxLogs=32 2025-01-08T10:45:09,979 INFO [RS:2;67f42ee69b4e:38323 {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=67f42ee69b4e%2C38323%2C1736333108259, suffix=, logDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/WALs/67f42ee69b4e,38323,1736333108259, archiveDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/oldWALs, maxLogs=32 2025-01-08T10:45:09,984 INFO [RS:0;67f42ee69b4e:45003 {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=67f42ee69b4e%2C45003%2C1736333108039, suffix=, logDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/WALs/67f42ee69b4e,45003,1736333108039, archiveDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/oldWALs, maxLogs=32 2025-01-08T10:45:10,003 DEBUG [RS:1;67f42ee69b4e:36283 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(617): When create output stream for /user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/WALs/67f42ee69b4e,36283,1736333108166/67f42ee69b4e%2C36283%2C1736333108166.1736333109982, exclude list is [], retry=0 2025-01-08T10:45:10,005 DEBUG [RS:2;67f42ee69b4e:38323 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(617): When create output stream for /user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/WALs/67f42ee69b4e,38323,1736333108259/67f42ee69b4e%2C38323%2C1736333108259.1736333109984, exclude list is [], retry=0 2025-01-08T10:45:10,008 DEBUG [RS:0;67f42ee69b4e:45003 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(617): When create output stream for /user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/WALs/67f42ee69b4e,45003,1736333108039/67f42ee69b4e%2C45003%2C1736333108039.1736333109986, exclude list is [], retry=0 2025-01-08T10:45:10,011 DEBUG [RS-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:43487,DS-96de2396-f7dd-4dd7-805c-c34c70b41204,DISK] 2025-01-08T10:45:10,012 DEBUG [RS-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45285,DS-5709618f-16cd-4409-9a05-d0991502eaf3,DISK] 2025-01-08T10:45:10,012 DEBUG [RS-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:44473,DS-687c546a-3440-42d4-b346-ee987087ce43,DISK] 2025-01-08T10:45:10,012 DEBUG [RS-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45285,DS-5709618f-16cd-4409-9a05-d0991502eaf3,DISK] 2025-01-08T10:45:10,012 DEBUG [RS-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:44473,DS-687c546a-3440-42d4-b346-ee987087ce43,DISK] 2025-01-08T10:45:10,013 DEBUG [RS-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:43487,DS-96de2396-f7dd-4dd7-805c-c34c70b41204,DISK] 2025-01-08T10:45:10,014 DEBUG [RS-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:43487,DS-96de2396-f7dd-4dd7-805c-c34c70b41204,DISK] 2025-01-08T10:45:10,015 DEBUG [RS-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:44473,DS-687c546a-3440-42d4-b346-ee987087ce43,DISK] 2025-01-08T10:45:10,015 DEBUG [RS-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45285,DS-5709618f-16cd-4409-9a05-d0991502eaf3,DISK] 2025-01-08T10:45:10,070 INFO [RS:1;67f42ee69b4e:36283 {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/WALs/67f42ee69b4e,36283,1736333108166/67f42ee69b4e%2C36283%2C1736333108166.1736333109982 2025-01-08T10:45:10,079 INFO [RS:2;67f42ee69b4e:38323 {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/WALs/67f42ee69b4e,38323,1736333108259/67f42ee69b4e%2C38323%2C1736333108259.1736333109984 2025-01-08T10:45:10,085 INFO [RS:0;67f42ee69b4e:45003 {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/WALs/67f42ee69b4e,45003,1736333108039/67f42ee69b4e%2C45003%2C1736333108039.1736333109986 2025-01-08T10:45:10,087 DEBUG [RS:1;67f42ee69b4e:36283 {}] wal.AbstractFSWAL(925): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:34377:34377),(127.0.0.1/127.0.0.1:40217:40217),(127.0.0.1/127.0.0.1:41419:41419)] 2025-01-08T10:45:10,087 DEBUG [RS:2;67f42ee69b4e:38323 {}] wal.AbstractFSWAL(925): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:40217:40217),(127.0.0.1/127.0.0.1:34377:34377),(127.0.0.1/127.0.0.1:41419:41419)] 2025-01-08T10:45:10,090 DEBUG [RS:0;67f42ee69b4e:45003 {}] wal.AbstractFSWAL(925): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:34377:34377),(127.0.0.1/127.0.0.1:41419:41419),(127.0.0.1/127.0.0.1:40217:40217)] 2025-01-08T10:45:10,158 DEBUG [67f42ee69b4e:34551 {}] assignment.AssignmentManager(2444): Processing assignQueue; systemServersCount=3, allServersCount=3 2025-01-08T10:45:10,162 DEBUG [67f42ee69b4e:34551 {}] balancer.BalancerClusterState(202): Hosts are {67f42ee69b4e=0} racks are {/default-rack=0} 2025-01-08T10:45:10,170 DEBUG [67f42ee69b4e:34551 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T10:45:10,170 DEBUG [67f42ee69b4e:34551 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T10:45:10,170 DEBUG [67f42ee69b4e:34551 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T10:45:10,170 INFO [67f42ee69b4e:34551 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T10:45:10,170 INFO [67f42ee69b4e:34551 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T10:45:10,170 INFO [67f42ee69b4e:34551 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T10:45:10,170 DEBUG [67f42ee69b4e:34551 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T10:45:10,175 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:10,181 INFO [PEWorker-2 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 67f42ee69b4e,36283,1736333108166, state=OPENING 2025-01-08T10:45:10,186 DEBUG [PEWorker-2 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2025-01-08T10:45:10,188 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:10,189 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:10,189 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T10:45:10,190 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:10,190 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:10,190 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T10:45:10,191 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T10:45:10,191 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T10:45:10,193 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE; OpenRegionProcedure 1588230740, server=67f42ee69b4e,36283,1736333108166}] 2025-01-08T10:45:10,368 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:10,370 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2025-01-08T10:45:10,372 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:37494, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2025-01-08T10:45:10,386 INFO [RS_OPEN_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(135): Open hbase:meta,,1.1588230740 2025-01-08T10:45:10,386 INFO [RS_OPEN_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2025-01-08T10:45:10,387 INFO [RS_OPEN_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor .meta 2025-01-08T10:45:10,390 INFO [RS_OPEN_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=67f42ee69b4e%2C36283%2C1736333108166.meta, suffix=.meta, logDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/WALs/67f42ee69b4e,36283,1736333108166, archiveDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/oldWALs, maxLogs=32 2025-01-08T10:45:10,407 DEBUG [RS_OPEN_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_META, pid=3}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(617): When create output stream for /user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/WALs/67f42ee69b4e,36283,1736333108166/67f42ee69b4e%2C36283%2C1736333108166.meta.1736333110392.meta, exclude list is [], retry=0 2025-01-08T10:45:10,412 DEBUG [RS-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:44473,DS-687c546a-3440-42d4-b346-ee987087ce43,DISK] 2025-01-08T10:45:10,412 DEBUG [RS-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:45285,DS-5709618f-16cd-4409-9a05-d0991502eaf3,DISK] 2025-01-08T10:45:10,412 DEBUG [RS-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:43487,DS-96de2396-f7dd-4dd7-805c-c34c70b41204,DISK] 2025-01-08T10:45:10,416 INFO [RS_OPEN_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/WALs/67f42ee69b4e,36283,1736333108166/67f42ee69b4e%2C36283%2C1736333108166.meta.1736333110392.meta 2025-01-08T10:45:10,416 DEBUG [RS_OPEN_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(925): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:40217:40217),(127.0.0.1/127.0.0.1:41419:41419),(127.0.0.1/127.0.0.1:34377:34377)] 2025-01-08T10:45:10,417 DEBUG [RS_OPEN_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7285): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2025-01-08T10:45:10,418 DEBUG [RS_OPEN_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7999): Registered coprocessor service: region=hbase:meta,,1 service=AccessControlService 2025-01-08T10:45:10,419 INFO [RS_OPEN_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:45:10,420 DEBUG [RS_OPEN_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2025-01-08T10:45:10,421 DEBUG [RS_OPEN_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7999): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2025-01-08T10:45:10,423 INFO [RS_OPEN_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(436): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2025-01-08T10:45:10,432 DEBUG [RS_OPEN_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2025-01-08T10:45:10,432 DEBUG [RS_OPEN_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:45:10,433 DEBUG [RS_OPEN_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7327): checking encryption for 1588230740 2025-01-08T10:45:10,433 DEBUG [RS_OPEN_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7330): checking classloading for 1588230740 2025-01-08T10:45:10,436 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2025-01-08T10:45:10,438 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2025-01-08T10:45:10,438 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:10,438 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2025-01-08T10:45:10,439 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2025-01-08T10:45:10,440 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2025-01-08T10:45:10,441 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:10,441 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2025-01-08T10:45:10,442 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2025-01-08T10:45:10,443 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2025-01-08T10:45:10,444 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:10,444 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2025-01-08T10:45:10,446 DEBUG [RS_OPEN_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/meta/1588230740 2025-01-08T10:45:10,449 DEBUG [RS_OPEN_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/meta/1588230740 2025-01-08T10:45:10,451 DEBUG [RS_OPEN_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (42.7 M)) instead. 2025-01-08T10:45:10,454 DEBUG [RS_OPEN_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1085): writing seq id for 1588230740 2025-01-08T10:45:10,456 INFO [RS_OPEN_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=64909839, jitterRate=-0.03276802599430084}}}, FlushLargeStoresPolicy{flushSizeLowerBound=44739242} 2025-01-08T10:45:10,460 DEBUG [RS_OPEN_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1001): Region open journal for 1588230740: 2025-01-08T10:45:10,470 INFO [RS_OPEN_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1736333110362 2025-01-08T10:45:10,484 DEBUG [RS_OPEN_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:meta,,1.1588230740 2025-01-08T10:45:10,485 INFO [RS_OPEN_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(164): Opened hbase:meta,,1.1588230740 2025-01-08T10:45:10,486 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:10,488 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 67f42ee69b4e,36283,1736333108166, state=OPEN 2025-01-08T10:45:10,491 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2025-01-08T10:45:10,491 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2025-01-08T10:45:10,491 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2025-01-08T10:45:10,491 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2025-01-08T10:45:10,491 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T10:45:10,491 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T10:45:10,491 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T10:45:10,491 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T10:45:10,497 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=3, resume processing ppid=2 2025-01-08T10:45:10,497 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=3, ppid=2, state=SUCCESS; OpenRegionProcedure 1588230740, server=67f42ee69b4e,36283,1736333108166 in 298 msec 2025-01-08T10:45:10,507 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=2, resume processing ppid=1 2025-01-08T10:45:10,507 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=2, ppid=1, state=SUCCESS; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 758 msec 2025-01-08T10:45:10,512 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=1, state=SUCCESS; InitMetaProcedure table=hbase:meta in 970 msec 2025-01-08T10:45:10,513 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] master.HMaster(1099): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1736333110513, completionTime=-1 2025-01-08T10:45:10,513 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] master.ServerManager(907): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2025-01-08T10:45:10,513 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] assignment.AssignmentManager(1747): Joining cluster... 2025-01-08T10:45:10,563 DEBUG [hconnection-0x5fcb6b98-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:45:10,566 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:37504, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:45:10,579 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] assignment.AssignmentManager(1759): Number of RegionServers=3 2025-01-08T10:45:10,579 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1736333170579 2025-01-08T10:45:10,579 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1736333230579 2025-01-08T10:45:10,579 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] assignment.AssignmentManager(1766): Joined the cluster in 65 msec 2025-01-08T10:45:10,600 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] balancer.RegionLocationFinder(172): Locality for region 1588230740 changed from -1.0 to 0.0, refreshing cache 2025-01-08T10:45:10,606 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=67f42ee69b4e,34551,1736333106957-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:10,607 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=67f42ee69b4e,34551,1736333106957-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:10,607 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=67f42ee69b4e,34551,1736333106957-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:10,608 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-67f42ee69b4e:34551, period=300000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:10,608 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:10,614 DEBUG [master/67f42ee69b4e:0.Chore.1 {}] janitor.CatalogJanitor(179): 2025-01-08T10:45:10,617 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] master.TableNamespaceManager(92): Namespace table not found. Creating... 2025-01-08T10:45:10,618 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] master.HMaster(2440): Client=null/null create 'hbase:namespace', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2025-01-08T10:45:10,625 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=hbase:namespace 2025-01-08T10:45:10,628 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T10:45:10,630 DEBUG [PEWorker-2 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:10,633 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T10:45:10,659 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741837_1013 (size=358) 2025-01-08T10:45:10,659 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741837_1013 (size=358) 2025-01-08T10:45:10,660 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741837_1013 (size=358) 2025-01-08T10:45:11,065 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 873898d3d33a2f6df797594f02de3737, NAME => 'hbase:namespace,,1736333110618.873898d3d33a2f6df797594f02de3737.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:namespace', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:45:11,080 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741838_1014 (size=42) 2025-01-08T10:45:11,080 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741838_1014 (size=42) 2025-01-08T10:45:11,082 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741838_1014 (size=42) 2025-01-08T10:45:11,083 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(894): Instantiated hbase:namespace,,1736333110618.873898d3d33a2f6df797594f02de3737.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:45:11,083 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1681): Closing 873898d3d33a2f6df797594f02de3737, disabling compactions & flushes 2025-01-08T10:45:11,083 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1703): Closing region hbase:namespace,,1736333110618.873898d3d33a2f6df797594f02de3737. 2025-01-08T10:45:11,083 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1736333110618.873898d3d33a2f6df797594f02de3737. 2025-01-08T10:45:11,083 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1736333110618.873898d3d33a2f6df797594f02de3737. after waiting 0 ms 2025-01-08T10:45:11,083 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1736333110618.873898d3d33a2f6df797594f02de3737. 2025-01-08T10:45:11,083 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1922): Closed hbase:namespace,,1736333110618.873898d3d33a2f6df797594f02de3737. 2025-01-08T10:45:11,083 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1635): Region close journal for 873898d3d33a2f6df797594f02de3737: 2025-01-08T10:45:11,086 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T10:45:11,093 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"hbase:namespace,,1736333110618.873898d3d33a2f6df797594f02de3737.","families":{"info":[{"qualifier":"regioninfo","vlen":41,"tag":[],"timestamp":"1736333111088"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736333111088"}]},"ts":"1736333111088"} 2025-01-08T10:45:11,123 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2025-01-08T10:45:11,125 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T10:45:11,129 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333111126"}]},"ts":"1736333111126"} 2025-01-08T10:45:11,135 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLING in hbase:meta 2025-01-08T10:45:11,140 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(202): Hosts are {67f42ee69b4e=0} racks are {/default-rack=0} 2025-01-08T10:45:11,141 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T10:45:11,141 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T10:45:11,141 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T10:45:11,141 INFO [PEWorker-2 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T10:45:11,141 INFO [PEWorker-2 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T10:45:11,141 INFO [PEWorker-2 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T10:45:11,141 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T10:45:11,143 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=873898d3d33a2f6df797594f02de3737, ASSIGN}] 2025-01-08T10:45:11,146 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=873898d3d33a2f6df797594f02de3737, ASSIGN 2025-01-08T10:45:11,149 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:namespace, region=873898d3d33a2f6df797594f02de3737, ASSIGN; state=OFFLINE, location=67f42ee69b4e,38323,1736333108259; forceNewPlan=false, retain=false 2025-01-08T10:45:11,300 INFO [67f42ee69b4e:34551 {}] balancer.BaseLoadBalancer(546): Reassigned 1 regions. 1 retained the pre-restart assignment. 2025-01-08T10:45:11,301 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=873898d3d33a2f6df797594f02de3737, regionState=OPENING, regionLocation=67f42ee69b4e,38323,1736333108259 2025-01-08T10:45:11,307 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE; OpenRegionProcedure 873898d3d33a2f6df797594f02de3737, server=67f42ee69b4e,38323,1736333108259}] 2025-01-08T10:45:11,461 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,38323,1736333108259 2025-01-08T10:45:11,462 DEBUG [RSProcedureDispatcher-pool-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2025-01-08T10:45:11,464 INFO [RS-EventLoopGroup-5-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:46588, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2025-01-08T10:45:11,471 INFO [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(135): Open hbase:namespace,,1736333110618.873898d3d33a2f6df797594f02de3737. 2025-01-08T10:45:11,471 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7285): Opening region: {ENCODED => 873898d3d33a2f6df797594f02de3737, NAME => 'hbase:namespace,,1736333110618.873898d3d33a2f6df797594f02de3737.', STARTKEY => '', ENDKEY => ''} 2025-01-08T10:45:11,471 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7999): Registered coprocessor service: region=hbase:namespace,,1736333110618.873898d3d33a2f6df797594f02de3737. service=AccessControlService 2025-01-08T10:45:11,471 INFO [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:45:11,472 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table namespace 873898d3d33a2f6df797594f02de3737 2025-01-08T10:45:11,472 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(894): Instantiated hbase:namespace,,1736333110618.873898d3d33a2f6df797594f02de3737.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:45:11,472 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7327): checking encryption for 873898d3d33a2f6df797594f02de3737 2025-01-08T10:45:11,472 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7330): checking classloading for 873898d3d33a2f6df797594f02de3737 2025-01-08T10:45:11,475 INFO [StoreOpener-873898d3d33a2f6df797594f02de3737-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 873898d3d33a2f6df797594f02de3737 2025-01-08T10:45:11,477 INFO [StoreOpener-873898d3d33a2f6df797594f02de3737-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 873898d3d33a2f6df797594f02de3737 columnFamilyName info 2025-01-08T10:45:11,477 DEBUG [StoreOpener-873898d3d33a2f6df797594f02de3737-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:11,478 INFO [StoreOpener-873898d3d33a2f6df797594f02de3737-1 {}] regionserver.HStore(327): Store=873898d3d33a2f6df797594f02de3737/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:45:11,480 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/namespace/873898d3d33a2f6df797594f02de3737 2025-01-08T10:45:11,484 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/namespace/873898d3d33a2f6df797594f02de3737 2025-01-08T10:45:11,489 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1085): writing seq id for 873898d3d33a2f6df797594f02de3737 2025-01-08T10:45:11,496 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/namespace/873898d3d33a2f6df797594f02de3737/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:45:11,497 INFO [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1102): Opened 873898d3d33a2f6df797594f02de3737; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=74063249, jitterRate=0.10362841188907623}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T10:45:11,499 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1001): Region open journal for 873898d3d33a2f6df797594f02de3737: 2025-01-08T10:45:11,502 INFO [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:namespace,,1736333110618.873898d3d33a2f6df797594f02de3737., pid=6, masterSystemTime=1736333111461 2025-01-08T10:45:11,507 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:namespace,,1736333110618.873898d3d33a2f6df797594f02de3737. 2025-01-08T10:45:11,508 INFO [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(164): Opened hbase:namespace,,1736333110618.873898d3d33a2f6df797594f02de3737. 2025-01-08T10:45:11,509 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=873898d3d33a2f6df797594f02de3737, regionState=OPEN, openSeqNum=2, regionLocation=67f42ee69b4e,38323,1736333108259 2025-01-08T10:45:11,520 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=6, resume processing ppid=5 2025-01-08T10:45:11,520 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=6, ppid=5, state=SUCCESS; OpenRegionProcedure 873898d3d33a2f6df797594f02de3737, server=67f42ee69b4e,38323,1736333108259 in 206 msec 2025-01-08T10:45:11,527 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=5, resume processing ppid=4 2025-01-08T10:45:11,527 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=5, ppid=4, state=SUCCESS; TransitRegionStateProcedure table=hbase:namespace, region=873898d3d33a2f6df797594f02de3737, ASSIGN in 377 msec 2025-01-08T10:45:11,529 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T10:45:11,529 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333111529"}]},"ts":"1736333111529"} 2025-01-08T10:45:11,533 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLED in hbase:meta 2025-01-08T10:45:11,537 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T10:45:11,540 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=4, state=SUCCESS; CreateTableProcedure table=hbase:namespace in 918 msec 2025-01-08T10:45:11,629 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] zookeeper.ZKUtil(113): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/namespace 2025-01-08T10:45:11,632 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/namespace 2025-01-08T10:45:11,632 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:11,632 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:11,632 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:11,632 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:11,660 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:45:11,661 INFO [RS-EventLoopGroup-5-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:46600, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:45:11,703 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=7, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=default 2025-01-08T10:45:11,721 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2025-01-08T10:45:11,732 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=7, state=SUCCESS; CreateNamespaceProcedure, namespace=default in 58 msec 2025-01-08T10:45:11,738 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=8, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=hbase 2025-01-08T10:45:11,757 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2025-01-08T10:45:11,763 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=8, state=SUCCESS; CreateNamespaceProcedure, namespace=hbase in 24 msec 2025-01-08T10:45:11,776 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/default 2025-01-08T10:45:11,779 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/hbase 2025-01-08T10:45:11,779 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] master.HMaster(1229): Master has completed initialization 3.285sec 2025-01-08T10:45:11,782 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2025-01-08T10:45:11,783 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2025-01-08T10:45:11,784 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2025-01-08T10:45:11,785 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2025-01-08T10:45:11,785 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2025-01-08T10:45:11,786 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=67f42ee69b4e,34551,1736333106957-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2025-01-08T10:45:11,786 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=67f42ee69b4e,34551,1736333106957-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2025-01-08T10:45:11,802 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] master.HMaster$4(2404): Client=null/null create 'hbase:acl', {NAME => 'l', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2025-01-08T10:45:11,804 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=9, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=hbase:acl 2025-01-08T10:45:11,807 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=hbase:acl execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T10:45:11,807 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:11,808 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] master.MasterRpcServices(713): Client=null/null procedure request for creating table: namespace: "hbase" qualifier: "acl" procId is: 9 2025-01-08T10:45:11,809 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=hbase:acl execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T10:45:11,813 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(149): Connect 0x067db5ed to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@5c4a797c 2025-01-08T10:45:11,814 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2025-01-08T10:45:11,814 WARN [Time-limited test {}] client.ZKConnectionRegistry(90): ZKConnectionRegistry is deprecated. See https://hbase.apache.org/book.html#client.rpcconnectionregistry 2025-01-08T10:45:11,821 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@738ef570, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:45:11,829 DEBUG [Time-limited test {}] nio.NioEventLoop(110): -Dio.netty.noKeySetOptimization: false 2025-01-08T10:45:11,829 DEBUG [Time-limited test {}] nio.NioEventLoop(111): -Dio.netty.selectorAutoRebuildThreshold: 512 2025-01-08T10:45:11,829 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741839_1015 (size=349) 2025-01-08T10:45:11,830 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741839_1015 (size=349) 2025-01-08T10:45:11,831 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741839_1015 (size=349) 2025-01-08T10:45:11,833 INFO [RegionOpenAndInit-hbase:acl-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 0c62df083bd10295a5e532128fcb0af8, NAME => 'hbase:acl,,1736333111799.0c62df083bd10295a5e532128fcb0af8.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:acl', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'l', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:45:11,863 DEBUG [hconnection-0x18c94e6b-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:45:11,876 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741840_1016 (size=36) 2025-01-08T10:45:11,876 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741840_1016 (size=36) 2025-01-08T10:45:11,877 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741840_1016 (size=36) 2025-01-08T10:45:11,881 DEBUG [RegionOpenAndInit-hbase:acl-pool-0 {}] regionserver.HRegion(894): Instantiated hbase:acl,,1736333111799.0c62df083bd10295a5e532128fcb0af8.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:45:11,881 DEBUG [RegionOpenAndInit-hbase:acl-pool-0 {}] regionserver.HRegion(1681): Closing 0c62df083bd10295a5e532128fcb0af8, disabling compactions & flushes 2025-01-08T10:45:11,881 INFO [RegionOpenAndInit-hbase:acl-pool-0 {}] regionserver.HRegion(1703): Closing region hbase:acl,,1736333111799.0c62df083bd10295a5e532128fcb0af8. 2025-01-08T10:45:11,881 DEBUG [RegionOpenAndInit-hbase:acl-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:acl,,1736333111799.0c62df083bd10295a5e532128fcb0af8. 2025-01-08T10:45:11,881 DEBUG [RegionOpenAndInit-hbase:acl-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on hbase:acl,,1736333111799.0c62df083bd10295a5e532128fcb0af8. after waiting 0 ms 2025-01-08T10:45:11,881 DEBUG [RegionOpenAndInit-hbase:acl-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region hbase:acl,,1736333111799.0c62df083bd10295a5e532128fcb0af8. 2025-01-08T10:45:11,882 INFO [RegionOpenAndInit-hbase:acl-pool-0 {}] regionserver.HRegion(1922): Closed hbase:acl,,1736333111799.0c62df083bd10295a5e532128fcb0af8. 2025-01-08T10:45:11,882 DEBUG [RegionOpenAndInit-hbase:acl-pool-0 {}] regionserver.HRegion(1635): Region close journal for 0c62df083bd10295a5e532128fcb0af8: 2025-01-08T10:45:11,882 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:37518, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:45:11,886 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1199): Minicluster is up; activeMaster=67f42ee69b4e,34551,1736333106957 2025-01-08T10:45:11,886 INFO [Time-limited test {}] hbase.HBaseTestingUtility(2790): Starting mini mapreduce cluster... 2025-01-08T10:45:11,886 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/test.cache.data in system properties and HBase conf 2025-01-08T10:45:11,886 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop.tmp.dir in system properties and HBase conf 2025-01-08T10:45:11,886 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop.log.dir in system properties and HBase conf 2025-01-08T10:45:11,886 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/mapreduce.cluster.local.dir in system properties and HBase conf 2025-01-08T10:45:11,886 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/mapreduce.cluster.temp.dir in system properties and HBase conf 2025-01-08T10:45:11,886 INFO [Time-limited test {}] hbase.HBaseTestingUtility(811): read short circuit is OFF 2025-01-08T10:45:11,887 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2025-01-08T10:45:11,887 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2025-01-08T10:45:11,887 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/yarn.nodemanager.log-dirs in system properties and HBase conf 2025-01-08T10:45:11,887 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2025-01-08T10:45:11,887 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2025-01-08T10:45:11,887 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2025-01-08T10:45:11,887 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2025-01-08T10:45:11,887 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/dfs.journalnode.edits.dir in system properties and HBase conf 2025-01-08T10:45:11,887 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2025-01-08T10:45:11,887 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/nfs.dump.dir in system properties and HBase conf 2025-01-08T10:45:11,888 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/java.io.tmpdir in system properties and HBase conf 2025-01-08T10:45:11,888 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/dfs.journalnode.edits.dir in system properties and HBase conf 2025-01-08T10:45:11,888 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2025-01-08T10:45:11,888 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2025-01-08T10:45:11,889 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=hbase:acl execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T10:45:11,889 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"hbase:acl,,1736333111799.0c62df083bd10295a5e532128fcb0af8.","families":{"info":[{"qualifier":"regioninfo","vlen":35,"tag":[],"timestamp":"1736333111889"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736333111889"}]},"ts":"1736333111889"} 2025-01-08T10:45:11,893 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2025-01-08T10:45:11,896 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=hbase:acl execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T10:45:11,897 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:acl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333111896"}]},"ts":"1736333111896"} 2025-01-08T10:45:11,900 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:acl, state=ENABLING in hbase:meta 2025-01-08T10:45:11,905 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(202): Hosts are {67f42ee69b4e=0} racks are {/default-rack=0} 2025-01-08T10:45:11,907 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T10:45:11,907 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T10:45:11,907 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T10:45:11,907 INFO [PEWorker-3 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T10:45:11,907 INFO [PEWorker-3 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T10:45:11,907 INFO [PEWorker-3 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T10:45:11,907 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T10:45:11,907 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:acl, region=0c62df083bd10295a5e532128fcb0af8, ASSIGN}] 2025-01-08T10:45:11,911 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:acl, region=0c62df083bd10295a5e532128fcb0af8, ASSIGN 2025-01-08T10:45:11,913 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:acl, region=0c62df083bd10295a5e532128fcb0af8, ASSIGN; state=OFFLINE, location=67f42ee69b4e,36283,1736333108166; forceNewPlan=false, retain=false 2025-01-08T10:45:11,916 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2025-01-08T10:45:11,989 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741841_1017 (size=592039) 2025-01-08T10:45:11,989 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741841_1017 (size=592039) 2025-01-08T10:45:11,991 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741841_1017 (size=592039) 2025-01-08T10:45:12,044 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741842_1018 (size=1663647) 2025-01-08T10:45:12,045 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741842_1018 (size=1663647) 2025-01-08T10:45:12,046 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741842_1018 (size=1663647) 2025-01-08T10:45:12,064 INFO [67f42ee69b4e:34551 {}] balancer.BaseLoadBalancer(546): Reassigned 1 regions. 1 retained the pre-restart assignment. 2025-01-08T10:45:12,064 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=10 updating hbase:meta row=0c62df083bd10295a5e532128fcb0af8, regionState=OPENING, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:12,075 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=11, ppid=10, state=RUNNABLE; OpenRegionProcedure 0c62df083bd10295a5e532128fcb0af8, server=67f42ee69b4e,36283,1736333108166}] 2025-01-08T10:45:12,116 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2025-01-08T10:45:12,247 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:12,293 INFO [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] handler.AssignRegionHandler(135): Open hbase:acl,,1736333111799.0c62df083bd10295a5e532128fcb0af8. 2025-01-08T10:45:12,293 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(7285): Opening region: {ENCODED => 0c62df083bd10295a5e532128fcb0af8, NAME => 'hbase:acl,,1736333111799.0c62df083bd10295a5e532128fcb0af8.', STARTKEY => '', ENDKEY => ''} 2025-01-08T10:45:12,294 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(7999): Registered coprocessor service: region=hbase:acl,,1736333111799.0c62df083bd10295a5e532128fcb0af8. service=AccessControlService 2025-01-08T10:45:12,295 INFO [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:45:12,295 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table acl 0c62df083bd10295a5e532128fcb0af8 2025-01-08T10:45:12,295 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(894): Instantiated hbase:acl,,1736333111799.0c62df083bd10295a5e532128fcb0af8.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:45:12,295 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(7327): checking encryption for 0c62df083bd10295a5e532128fcb0af8 2025-01-08T10:45:12,296 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(7330): checking classloading for 0c62df083bd10295a5e532128fcb0af8 2025-01-08T10:45:12,299 INFO [StoreOpener-0c62df083bd10295a5e532128fcb0af8-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family l of region 0c62df083bd10295a5e532128fcb0af8 2025-01-08T10:45:12,302 INFO [StoreOpener-0c62df083bd10295a5e532128fcb0af8-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 0c62df083bd10295a5e532128fcb0af8 columnFamilyName l 2025-01-08T10:45:12,302 DEBUG [StoreOpener-0c62df083bd10295a5e532128fcb0af8-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:12,303 INFO [StoreOpener-0c62df083bd10295a5e532128fcb0af8-1 {}] regionserver.HStore(327): Store=0c62df083bd10295a5e532128fcb0af8/l, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:45:12,305 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/acl/0c62df083bd10295a5e532128fcb0af8 2025-01-08T10:45:12,306 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/acl/0c62df083bd10295a5e532128fcb0af8 2025-01-08T10:45:12,310 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(1085): writing seq id for 0c62df083bd10295a5e532128fcb0af8 2025-01-08T10:45:12,314 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/acl/0c62df083bd10295a5e532128fcb0af8/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:45:12,315 INFO [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(1102): Opened 0c62df083bd10295a5e532128fcb0af8; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=73537608, jitterRate=0.09579575061798096}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T10:45:12,318 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(1001): Region open journal for 0c62df083bd10295a5e532128fcb0af8: 2025-01-08T10:45:12,320 INFO [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:acl,,1736333111799.0c62df083bd10295a5e532128fcb0af8., pid=11, masterSystemTime=1736333112247 2025-01-08T10:45:12,323 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:acl,,1736333111799.0c62df083bd10295a5e532128fcb0af8. 2025-01-08T10:45:12,323 INFO [RS_OPEN_PRIORITY_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] handler.AssignRegionHandler(164): Opened hbase:acl,,1736333111799.0c62df083bd10295a5e532128fcb0af8. 2025-01-08T10:45:12,324 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=10 updating hbase:meta row=0c62df083bd10295a5e532128fcb0af8, regionState=OPEN, openSeqNum=2, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:12,333 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=11, resume processing ppid=10 2025-01-08T10:45:12,333 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=11, ppid=10, state=SUCCESS; OpenRegionProcedure 0c62df083bd10295a5e532128fcb0af8, server=67f42ee69b4e,36283,1736333108166 in 254 msec 2025-01-08T10:45:12,339 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=10, resume processing ppid=9 2025-01-08T10:45:12,339 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=10, ppid=9, state=SUCCESS; TransitRegionStateProcedure table=hbase:acl, region=0c62df083bd10295a5e532128fcb0af8, ASSIGN in 426 msec 2025-01-08T10:45:12,341 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=hbase:acl execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T10:45:12,341 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:acl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333112341"}]},"ts":"1736333112341"} 2025-01-08T10:45:12,345 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:acl, state=ENABLED in hbase:meta 2025-01-08T10:45:12,353 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=hbase:acl execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T10:45:12,357 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=9, state=SUCCESS; CreateTableProcedure table=hbase:acl in 551 msec 2025-01-08T10:45:12,417 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2025-01-08T10:45:12,417 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: hbase:acl, procId: 9 completed 2025-01-08T10:45:12,421 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster {}] master.HMaster(1332): Balancer post startup initialization complete, took 0 seconds 2025-01-08T10:45:12,422 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2025-01-08T10:45:12,423 INFO [master/67f42ee69b4e:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=67f42ee69b4e,34551,1736333106957-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2025-01-08T10:45:14,408 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T10:45:14,572 WARN [Thread-399 {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T10:45:14,926 WARN [Time-limited test {}] servlet.GuiceFilter(102): Multiple Servlet injectors detected. This is a warning indicating that you have more than one GuiceFilter running in your web application. If this is deliberate, you may safely ignore this message. If this is NOT deliberate however, your application may not work as expected. 2025-01-08T10:45:14,927 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2025-01-08T10:45:14,928 INFO [Thread-399 {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2025-01-08T10:45:15,081 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2025-01-08T10:45:15,081 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2025-01-08T10:45:15,081 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2025-01-08T10:45:15,083 INFO [Thread-399 {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2025-01-08T10:45:15,083 INFO [Thread-399 {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2025-01-08T10:45:15,083 INFO [Thread-399 {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2025-01-08T10:45:15,086 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T10:45:15,096 INFO [Thread-399 {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@431bcc59{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop.log.dir/,AVAILABLE} 2025-01-08T10:45:15,096 INFO [Thread-399 {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3a64ffa4{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/static,AVAILABLE} 2025-01-08T10:45:15,115 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@5d5a85fc{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop.log.dir/,AVAILABLE} 2025-01-08T10:45:15,116 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@fc1b748{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/static,AVAILABLE} 2025-01-08T10:45:15,351 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(113): Registering org.apache.hadoop.yarn.server.resourcemanager.webapp.JAXBContextResolver as a provider class 2025-01-08T10:45:15,351 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(116): Registering org.apache.hadoop.yarn.server.resourcemanager.webapp.RMWebServices as a root resource class 2025-01-08T10:45:15,352 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(113): Registering org.apache.hadoop.yarn.webapp.GenericExceptionHandler as a provider class 2025-01-08T10:45:15,354 INFO [Time-limited test {}] application.WebApplicationImpl(815): Initiating Jersey application, version 'Jersey: 1.19.4 05/24/2017 03:20 PM' 2025-01-08T10:45:15,555 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(168): Binding org.apache.hadoop.yarn.server.resourcemanager.webapp.JAXBContextResolver to GuiceManagedComponentProvider with the scope "Singleton" 2025-01-08T10:45:15,989 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T10:45:16,098 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:namespace' 2025-01-08T10:45:16,100 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2025-01-08T10:45:16,101 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:acl' 2025-01-08T10:45:16,179 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(168): Binding org.apache.hadoop.yarn.webapp.GenericExceptionHandler to GuiceManagedComponentProvider with the scope "Singleton" 2025-01-08T10:45:16,577 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(168): Binding org.apache.hadoop.yarn.server.resourcemanager.webapp.RMWebServices to GuiceManagedComponentProvider with the scope "Singleton" 2025-01-08T10:45:16,603 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@7c1fb4a6{cluster,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/java.io.tmpdir/jetty-localhost-36391-hadoop-yarn-common-3_4_1_jar-_-any-7018933113267787803/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/cluster} 2025-01-08T10:45:16,603 INFO [Thread-399 {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@7b7fab8{jobhistory,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/java.io.tmpdir/jetty-localhost-42677-hadoop-yarn-common-3_4_1_jar-_-any-14872496750938547664/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/jobhistory} 2025-01-08T10:45:16,604 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@3aa3cd9c{HTTP/1.1, (http/1.1)}{localhost:36391} 2025-01-08T10:45:16,604 INFO [Thread-399 {}] server.AbstractConnector(333): Started ServerConnector@7c1e8a21{HTTP/1.1, (http/1.1)}{localhost:42677} 2025-01-08T10:45:16,604 INFO [Time-limited test {}] server.Server(415): Started @18844ms 2025-01-08T10:45:16,604 INFO [Thread-399 {}] server.Server(415): Started @18844ms 2025-01-08T10:45:16,864 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741843_1019 (size=5) 2025-01-08T10:45:16,865 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741843_1019 (size=5) 2025-01-08T10:45:16,866 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741843_1019 (size=5) 2025-01-08T10:45:17,596 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:45:17,596 WARN [Time-limited test {}] tracker.NMLogAggregationStatusTracker(95): Log Aggregation is disabled.So is the LogAggregationStatusTracker. 2025-01-08T10:45:17,596 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.security.access.AccessController Metrics about HBase RegionObservers 2025-01-08T10:45:17,597 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_master_table_store 2025-01-08T10:45:17,597 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_master_table_store Metrics about Tables on a single HBase RegionServer 2025-01-08T10:45:17,599 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=Coprocessor.RegionServer.CP_org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:45:17,599 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=Coprocessor.RegionServer.CP_org.apache.hadoop.hbase.security.access.AccessController Metrics about HBase RegionServerObservers 2025-01-08T10:45:17,600 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_meta 2025-01-08T10:45:17,600 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_meta Metrics about Tables on a single HBase RegionServer 2025-01-08T10:45:17,601 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_namespace 2025-01-08T10:45:17,601 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_namespace Metrics about Tables on a single HBase RegionServer 2025-01-08T10:45:17,602 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T10:45:17,603 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_acl 2025-01-08T10:45:17,603 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_acl Metrics about Tables on a single HBase RegionServer 2025-01-08T10:45:17,605 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: Master,sub=Coprocessor.Master.CP_org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:45:17,605 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering Master,sub=Coprocessor.Master.CP_org.apache.hadoop.hbase.security.access.AccessController Metrics about HBase MasterObservers 2025-01-08T10:45:17,605 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: Master,sub=Coprocessor.Master.CP_org.apache.hadoop.hbase.security.access.SecureTestUtil$MasterSyncObserver 2025-01-08T10:45:17,605 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering Master,sub=Coprocessor.Master.CP_org.apache.hadoop.hbase.security.access.SecureTestUtil$MasterSyncObserver Metrics about HBase MasterObservers 2025-01-08T10:45:17,606 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2025-01-08T10:45:17,606 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint Metrics about HBase RegionObservers 2025-01-08T10:45:17,641 WARN [Time-limited test {}] servlet.GuiceFilter(102): Multiple Servlet injectors detected. This is a warning indicating that you have more than one GuiceFilter running in your web application. If this is deliberate, you may safely ignore this message. If this is NOT deliberate however, your application may not work as expected. 2025-01-08T10:45:17,642 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2025-01-08T10:45:17,644 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2025-01-08T10:45:17,644 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2025-01-08T10:45:17,644 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2025-01-08T10:45:17,646 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T10:45:17,647 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@10e17038{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop.log.dir/,AVAILABLE} 2025-01-08T10:45:17,647 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@47e31f8f{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/static,AVAILABLE} 2025-01-08T10:45:17,717 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(116): Registering org.apache.hadoop.yarn.server.nodemanager.webapp.NMWebServices as a root resource class 2025-01-08T10:45:17,717 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(113): Registering org.apache.hadoop.yarn.webapp.GenericExceptionHandler as a provider class 2025-01-08T10:45:17,717 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(113): Registering org.apache.hadoop.yarn.server.nodemanager.webapp.JAXBContextResolver as a provider class 2025-01-08T10:45:17,717 INFO [Time-limited test {}] application.WebApplicationImpl(815): Initiating Jersey application, version 'Jersey: 1.19.4 05/24/2017 03:20 PM' 2025-01-08T10:45:17,726 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(168): Binding org.apache.hadoop.yarn.server.nodemanager.webapp.JAXBContextResolver to GuiceManagedComponentProvider with the scope "Singleton" 2025-01-08T10:45:17,747 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(168): Binding org.apache.hadoop.yarn.webapp.GenericExceptionHandler to GuiceManagedComponentProvider with the scope "Singleton" 2025-01-08T10:45:17,856 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(168): Binding org.apache.hadoop.yarn.server.nodemanager.webapp.NMWebServices to GuiceManagedComponentProvider with the scope "Singleton" 2025-01-08T10:45:17,867 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@f3ffbde{node,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/java.io.tmpdir/jetty-localhost-35355-hadoop-yarn-common-3_4_1_jar-_-any-10322108900160842982/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/node} 2025-01-08T10:45:17,867 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@61329307{HTTP/1.1, (http/1.1)}{localhost:35355} 2025-01-08T10:45:17,868 INFO [Time-limited test {}] server.Server(415): Started @20107ms 2025-01-08T10:45:18,047 WARN [Time-limited test {}] tracker.NMLogAggregationStatusTracker(95): Log Aggregation is disabled.So is the LogAggregationStatusTracker. 2025-01-08T10:45:18,049 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T10:45:18,061 WARN [Time-limited test {}] servlet.GuiceFilter(102): Multiple Servlet injectors detected. This is a warning indicating that you have more than one GuiceFilter running in your web application. If this is deliberate, you may safely ignore this message. If this is NOT deliberate however, your application may not work as expected. 2025-01-08T10:45:18,061 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2025-01-08T10:45:18,063 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2025-01-08T10:45:18,063 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2025-01-08T10:45:18,063 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2025-01-08T10:45:18,064 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T10:45:18,064 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@5f820e1d{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop.log.dir/,AVAILABLE} 2025-01-08T10:45:18,065 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6868b947{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/static,AVAILABLE} 2025-01-08T10:45:18,122 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(116): Registering org.apache.hadoop.yarn.server.nodemanager.webapp.NMWebServices as a root resource class 2025-01-08T10:45:18,122 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(113): Registering org.apache.hadoop.yarn.webapp.GenericExceptionHandler as a provider class 2025-01-08T10:45:18,122 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(113): Registering org.apache.hadoop.yarn.server.nodemanager.webapp.JAXBContextResolver as a provider class 2025-01-08T10:45:18,122 INFO [Time-limited test {}] application.WebApplicationImpl(815): Initiating Jersey application, version 'Jersey: 1.19.4 05/24/2017 03:20 PM' 2025-01-08T10:45:18,132 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(168): Binding org.apache.hadoop.yarn.server.nodemanager.webapp.JAXBContextResolver to GuiceManagedComponentProvider with the scope "Singleton" 2025-01-08T10:45:18,139 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(168): Binding org.apache.hadoop.yarn.webapp.GenericExceptionHandler to GuiceManagedComponentProvider with the scope "Singleton" 2025-01-08T10:45:18,254 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(168): Binding org.apache.hadoop.yarn.server.nodemanager.webapp.NMWebServices to GuiceManagedComponentProvider with the scope "Singleton" 2025-01-08T10:45:18,260 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@2d6c849d{node,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/java.io.tmpdir/jetty-localhost-34777-hadoop-yarn-common-3_4_1_jar-_-any-16054929641418138832/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/node} 2025-01-08T10:45:18,261 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@759bc9dd{HTTP/1.1, (http/1.1)}{localhost:34777} 2025-01-08T10:45:18,261 INFO [Time-limited test {}] server.Server(415): Started @20501ms 2025-01-08T10:45:18,290 INFO [Time-limited test {}] hbase.HBaseTestingUtility(2825): Mini mapreduce cluster started 2025-01-08T10:45:18,292 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [30,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T10:45:18,321 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: snapshot.TestMobSecureExportSnapshot#testExportWithTargetName Thread=718, OpenFileDescriptor=781, MaxFileDescriptor=1048576, SystemLoadAverage=735, ProcessCount=19, AvailableMemoryMB=982 2025-01-08T10:45:18,321 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=718 is superior to 500 2025-01-08T10:45:18,333 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2025-01-08T10:45:18,335 INFO [RS-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:48402, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2025-01-08T10:45:18,340 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.2 create 'testtb-testExportWithTargetName', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T10:45:18,342 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=12, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testExportWithTargetName 2025-01-08T10:45:18,344 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=12, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportWithTargetName execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T10:45:18,344 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "testtb-testExportWithTargetName" procId is: 12 2025-01-08T10:45:18,346 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=12 2025-01-08T10:45:18,346 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=12, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testExportWithTargetName execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T10:45:18,361 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741844_1020 (size=442) 2025-01-08T10:45:18,361 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741844_1020 (size=442) 2025-01-08T10:45:18,362 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741844_1020 (size=442) 2025-01-08T10:45:18,365 INFO [RegionOpenAndInit-testtb-testExportWithTargetName-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => a23b54a820be126874ca84971483afef, NAME => 'testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testtb-testExportWithTargetName', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:45:18,365 INFO [RegionOpenAndInit-testtb-testExportWithTargetName-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => 6ac2ef231908ea1ef7af5e35d2ad7d34, NAME => 'testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testtb-testExportWithTargetName', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:45:18,377 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741846_1022 (size=67) 2025-01-08T10:45:18,377 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741845_1021 (size=67) 2025-01-08T10:45:18,377 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741846_1022 (size=67) 2025-01-08T10:45:18,378 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741845_1021 (size=67) 2025-01-08T10:45:18,378 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741845_1021 (size=67) 2025-01-08T10:45:18,378 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741846_1022 (size=67) 2025-01-08T10:45:18,379 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:45:18,379 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-1 {}] regionserver.HRegion(1681): Closing 6ac2ef231908ea1ef7af5e35d2ad7d34, disabling compactions & flushes 2025-01-08T10:45:18,379 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:45:18,380 INFO [RegionOpenAndInit-testtb-testExportWithTargetName-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34. 2025-01-08T10:45:18,380 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34. 2025-01-08T10:45:18,380 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34. after waiting 0 ms 2025-01-08T10:45:18,380 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-0 {}] regionserver.HRegion(1681): Closing a23b54a820be126874ca84971483afef, disabling compactions & flushes 2025-01-08T10:45:18,380 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34. 2025-01-08T10:45:18,380 INFO [RegionOpenAndInit-testtb-testExportWithTargetName-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef. 2025-01-08T10:45:18,380 INFO [RegionOpenAndInit-testtb-testExportWithTargetName-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34. 2025-01-08T10:45:18,380 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef. 2025-01-08T10:45:18,380 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-1 {}] regionserver.HRegion(1635): Region close journal for 6ac2ef231908ea1ef7af5e35d2ad7d34: 2025-01-08T10:45:18,380 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef. after waiting 0 ms 2025-01-08T10:45:18,380 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef. 2025-01-08T10:45:18,380 INFO [RegionOpenAndInit-testtb-testExportWithTargetName-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef. 2025-01-08T10:45:18,380 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-0 {}] regionserver.HRegion(1635): Region close journal for a23b54a820be126874ca84971483afef: 2025-01-08T10:45:18,382 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=12, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testExportWithTargetName execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T10:45:18,382 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34.","families":{"info":[{"qualifier":"regioninfo","vlen":66,"tag":[],"timestamp":"1736333118382"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736333118382"}]},"ts":"1736333118382"} 2025-01-08T10:45:18,382 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef.","families":{"info":[{"qualifier":"regioninfo","vlen":66,"tag":[],"timestamp":"1736333118382"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736333118382"}]},"ts":"1736333118382"} 2025-01-08T10:45:18,414 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T10:45:18,415 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=12, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testExportWithTargetName execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T10:45:18,416 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithTargetName","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333118416"}]},"ts":"1736333118416"} 2025-01-08T10:45:18,418 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithTargetName, state=ENABLING in hbase:meta 2025-01-08T10:45:18,424 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(202): Hosts are {67f42ee69b4e=0} racks are {/default-rack=0} 2025-01-08T10:45:18,426 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T10:45:18,426 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T10:45:18,426 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T10:45:18,426 INFO [PEWorker-4 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T10:45:18,427 INFO [PEWorker-4 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T10:45:18,427 INFO [PEWorker-4 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T10:45:18,427 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T10:45:18,427 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=13, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=a23b54a820be126874ca84971483afef, ASSIGN}, {pid=14, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=6ac2ef231908ea1ef7af5e35d2ad7d34, ASSIGN}] 2025-01-08T10:45:18,429 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=14, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=6ac2ef231908ea1ef7af5e35d2ad7d34, ASSIGN 2025-01-08T10:45:18,429 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=13, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=a23b54a820be126874ca84971483afef, ASSIGN 2025-01-08T10:45:18,431 INFO [PEWorker-1 {}] assignment.TransitRegionStateProcedure(264): Starting pid=14, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=6ac2ef231908ea1ef7af5e35d2ad7d34, ASSIGN; state=OFFLINE, location=67f42ee69b4e,36283,1736333108166; forceNewPlan=false, retain=false 2025-01-08T10:45:18,431 INFO [PEWorker-5 {}] assignment.TransitRegionStateProcedure(264): Starting pid=13, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=a23b54a820be126874ca84971483afef, ASSIGN; state=OFFLINE, location=67f42ee69b4e,45003,1736333108039; forceNewPlan=false, retain=false 2025-01-08T10:45:18,449 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=12 2025-01-08T10:45:18,581 INFO [67f42ee69b4e:34551 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T10:45:18,582 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=14 updating hbase:meta row=6ac2ef231908ea1ef7af5e35d2ad7d34, regionState=OPENING, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:18,582 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=13 updating hbase:meta row=a23b54a820be126874ca84971483afef, regionState=OPENING, regionLocation=67f42ee69b4e,45003,1736333108039 2025-01-08T10:45:18,585 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=15, ppid=14, state=RUNNABLE; OpenRegionProcedure 6ac2ef231908ea1ef7af5e35d2ad7d34, server=67f42ee69b4e,36283,1736333108166}] 2025-01-08T10:45:18,587 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=16, ppid=13, state=RUNNABLE; OpenRegionProcedure a23b54a820be126874ca84971483afef, server=67f42ee69b4e,45003,1736333108039}] 2025-01-08T10:45:18,651 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=12 2025-01-08T10:45:18,739 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:18,741 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:45:18,741 DEBUG [RSProcedureDispatcher-pool-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2025-01-08T10:45:18,747 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] handler.AssignRegionHandler(135): Open testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34. 2025-01-08T10:45:18,747 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(7285): Opening region: {ENCODED => 6ac2ef231908ea1ef7af5e35d2ad7d34, NAME => 'testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T10:45:18,747 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34. service=AccessControlService 2025-01-08T10:45:18,747 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:45:18,748 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportWithTargetName 6ac2ef231908ea1ef7af5e35d2ad7d34 2025-01-08T10:45:18,748 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(894): Instantiated testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:45:18,748 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(7327): checking encryption for 6ac2ef231908ea1ef7af5e35d2ad7d34 2025-01-08T10:45:18,748 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(7330): checking classloading for 6ac2ef231908ea1ef7af5e35d2ad7d34 2025-01-08T10:45:18,752 INFO [StoreOpener-6ac2ef231908ea1ef7af5e35d2ad7d34-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 6ac2ef231908ea1ef7af5e35d2ad7d34 2025-01-08T10:45:18,759 INFO [StoreOpener-6ac2ef231908ea1ef7af5e35d2ad7d34-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 6ac2ef231908ea1ef7af5e35d2ad7d34 columnFamilyName cf 2025-01-08T10:45:18,761 INFO [RS-EventLoopGroup-3-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:35304, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2025-01-08T10:45:18,765 DEBUG [StoreOpener-6ac2ef231908ea1ef7af5e35d2ad7d34-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:18,766 INFO [StoreOpener-6ac2ef231908ea1ef7af5e35d2ad7d34-1 {}] regionserver.HStore(327): Store=6ac2ef231908ea1ef7af5e35d2ad7d34/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:45:18,768 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] handler.AssignRegionHandler(135): Open testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef. 2025-01-08T10:45:18,768 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/6ac2ef231908ea1ef7af5e35d2ad7d34 2025-01-08T10:45:18,769 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(7285): Opening region: {ENCODED => a23b54a820be126874ca84971483afef, NAME => 'testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T10:45:18,769 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef. service=AccessControlService 2025-01-08T10:45:18,769 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/6ac2ef231908ea1ef7af5e35d2ad7d34 2025-01-08T10:45:18,769 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:45:18,770 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportWithTargetName a23b54a820be126874ca84971483afef 2025-01-08T10:45:18,770 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(894): Instantiated testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:45:18,770 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(7327): checking encryption for a23b54a820be126874ca84971483afef 2025-01-08T10:45:18,770 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(7330): checking classloading for a23b54a820be126874ca84971483afef 2025-01-08T10:45:18,772 INFO [StoreOpener-a23b54a820be126874ca84971483afef-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region a23b54a820be126874ca84971483afef 2025-01-08T10:45:18,773 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(1085): writing seq id for 6ac2ef231908ea1ef7af5e35d2ad7d34 2025-01-08T10:45:18,775 INFO [StoreOpener-a23b54a820be126874ca84971483afef-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region a23b54a820be126874ca84971483afef columnFamilyName cf 2025-01-08T10:45:18,776 DEBUG [StoreOpener-a23b54a820be126874ca84971483afef-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:18,777 INFO [StoreOpener-a23b54a820be126874ca84971483afef-1 {}] regionserver.HStore(327): Store=a23b54a820be126874ca84971483afef/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:45:18,778 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/a23b54a820be126874ca84971483afef 2025-01-08T10:45:18,779 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/a23b54a820be126874ca84971483afef 2025-01-08T10:45:18,781 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/6ac2ef231908ea1ef7af5e35d2ad7d34/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:45:18,781 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(1102): Opened 6ac2ef231908ea1ef7af5e35d2ad7d34; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=63707299, jitterRate=-0.0506872683763504}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T10:45:18,782 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(1085): writing seq id for a23b54a820be126874ca84971483afef 2025-01-08T10:45:18,783 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(1001): Region open journal for 6ac2ef231908ea1ef7af5e35d2ad7d34: 2025-01-08T10:45:18,785 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34., pid=15, masterSystemTime=1736333118739 2025-01-08T10:45:18,785 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/a23b54a820be126874ca84971483afef/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:45:18,786 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(1102): Opened a23b54a820be126874ca84971483afef; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=74410655, jitterRate=0.10880516469478607}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T10:45:18,786 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(1001): Region open journal for a23b54a820be126874ca84971483afef: 2025-01-08T10:45:18,787 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef., pid=16, masterSystemTime=1736333118741 2025-01-08T10:45:18,788 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34. 2025-01-08T10:45:18,788 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] handler.AssignRegionHandler(164): Opened testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34. 2025-01-08T10:45:18,789 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=14 updating hbase:meta row=6ac2ef231908ea1ef7af5e35d2ad7d34, regionState=OPEN, openSeqNum=2, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:18,790 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef. 2025-01-08T10:45:18,790 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] handler.AssignRegionHandler(164): Opened testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef. 2025-01-08T10:45:18,791 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=13 updating hbase:meta row=a23b54a820be126874ca84971483afef, regionState=OPEN, openSeqNum=2, regionLocation=67f42ee69b4e,45003,1736333108039 2025-01-08T10:45:18,797 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=15, resume processing ppid=14 2025-01-08T10:45:18,797 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=15, ppid=14, state=SUCCESS; OpenRegionProcedure 6ac2ef231908ea1ef7af5e35d2ad7d34, server=67f42ee69b4e,36283,1736333108166 in 207 msec 2025-01-08T10:45:18,799 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=14, ppid=12, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=6ac2ef231908ea1ef7af5e35d2ad7d34, ASSIGN in 370 msec 2025-01-08T10:45:18,799 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=16, resume processing ppid=13 2025-01-08T10:45:18,800 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=16, ppid=13, state=SUCCESS; OpenRegionProcedure a23b54a820be126874ca84971483afef, server=67f42ee69b4e,45003,1736333108039 in 208 msec 2025-01-08T10:45:18,802 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=13, resume processing ppid=12 2025-01-08T10:45:18,802 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=13, ppid=12, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=a23b54a820be126874ca84971483afef, ASSIGN in 372 msec 2025-01-08T10:45:18,804 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=12, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testExportWithTargetName execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T10:45:18,804 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithTargetName","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333118804"}]},"ts":"1736333118804"} 2025-01-08T10:45:18,806 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithTargetName, state=ENABLED in hbase:meta 2025-01-08T10:45:18,810 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=12, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportWithTargetName execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T10:45:18,814 DEBUG [PEWorker-5 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testExportWithTargetName jenkins: RWXCA 2025-01-08T10:45:18,828 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36283 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportWithTargetName], kv [jenkins: RWXCA] 2025-01-08T10:45:18,851 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/acl 2025-01-08T10:45:18,851 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/acl 2025-01-08T10:45:18,851 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/acl 2025-01-08T10:45:18,851 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/acl 2025-01-08T10:45:18,851 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:18,851 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:18,851 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:18,852 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:45:18,861 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithTargetName with data PBUF\x0AE\x0A\x07jenkins\x12:\x08\x03"6\x0A*\x0A\x07default\x12\x1Ftesttb-testExportWithTargetName \x00 \x01 \x02 \x03 \x04 2025-01-08T10:45:18,861 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithTargetName with data PBUF\x0AE\x0A\x07jenkins\x12:\x08\x03"6\x0A*\x0A\x07default\x12\x1Ftesttb-testExportWithTargetName \x00 \x01 \x02 \x03 \x04 2025-01-08T10:45:18,861 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithTargetName with data PBUF\x0AE\x0A\x07jenkins\x12:\x08\x03"6\x0A*\x0A\x07default\x12\x1Ftesttb-testExportWithTargetName \x00 \x01 \x02 \x03 \x04 2025-01-08T10:45:18,861 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithTargetName with data PBUF\x0AE\x0A\x07jenkins\x12:\x08\x03"6\x0A*\x0A\x07default\x12\x1Ftesttb-testExportWithTargetName \x00 \x01 \x02 \x03 \x04 2025-01-08T10:45:18,865 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=12, state=SUCCESS; CreateTableProcedure table=testtb-testExportWithTargetName in 522 msec 2025-01-08T10:45:18,953 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=12 2025-01-08T10:45:18,953 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testExportWithTargetName, procId: 12 completed 2025-01-08T10:45:18,958 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportWithTargetName 2025-01-08T10:45:18,960 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef. 2025-01-08T10:45:18,961 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T10:45:18,988 DEBUG [hconnection-0x18c94e6b-shared-pool-4 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:45:18,990 INFO [RS-EventLoopGroup-3-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:35310, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:45:19,005 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.2 snapshot request for:{ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } 2025-01-08T10:45:19,006 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736333119006 (current time:1736333119006). 2025-01-08T10:45:19,006 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T10:45:19,006 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot emptySnaptb0-testExportWithTargetName VERSION not specified, setting to 2 2025-01-08T10:45:19,006 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T10:45:19,008 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x7acdb912 to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@416cf16 2025-01-08T10:45:19,013 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2c8ee579, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:45:19,015 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:45:19,016 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:42476, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:45:19,019 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x7acdb912 to 127.0.0.1:52367 2025-01-08T10:45:19,019 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:45:19,021 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x19ec304c to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@692be2a9 2025-01-08T10:45:19,026 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@363be6b0, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:45:19,032 DEBUG [hconnection-0x4f5c26fd-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:45:19,033 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:42484, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:45:19,038 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x19ec304c to 127.0.0.1:52367 2025-01-08T10:45:19,039 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:45:19,039 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportWithTargetName], kv [jenkins: RWXCA] 2025-01-08T10:45:19,055 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T10:45:19,062 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=17, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } 2025-01-08T10:45:19,063 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 }, snapshot procedure id = 17 2025-01-08T10:45:19,064 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T10:45:19,066 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=17 2025-01-08T10:45:19,072 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T10:45:19,084 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T10:45:19,096 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741847_1023 (size=167) 2025-01-08T10:45:19,097 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741847_1023 (size=167) 2025-01-08T10:45:19,098 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741847_1023 (size=167) 2025-01-08T10:45:19,100 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T10:45:19,103 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=18, ppid=17, state=RUNNABLE; SnapshotRegionProcedure a23b54a820be126874ca84971483afef}, {pid=19, ppid=17, state=RUNNABLE; SnapshotRegionProcedure 6ac2ef231908ea1ef7af5e35d2ad7d34}] 2025-01-08T10:45:19,107 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=18, ppid=17, state=RUNNABLE; SnapshotRegionProcedure a23b54a820be126874ca84971483afef 2025-01-08T10:45:19,107 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=19, ppid=17, state=RUNNABLE; SnapshotRegionProcedure 6ac2ef231908ea1ef7af5e35d2ad7d34 2025-01-08T10:45:19,168 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=17 2025-01-08T10:45:19,263 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:45:19,263 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:19,265 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=36283 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=19 2025-01-08T10:45:19,265 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=45003 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=18 2025-01-08T10:45:19,266 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=18}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef. 2025-01-08T10:45:19,266 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=19}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34. 2025-01-08T10:45:19,268 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=18}] regionserver.HRegion(2538): Flush status journal for a23b54a820be126874ca84971483afef: 2025-01-08T10:45:19,268 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=19}] regionserver.HRegion(2538): Flush status journal for 6ac2ef231908ea1ef7af5e35d2ad7d34: 2025-01-08T10:45:19,268 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=19}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34. for emptySnaptb0-testExportWithTargetName completed. 2025-01-08T10:45:19,268 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=18}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef. for emptySnaptb0-testExportWithTargetName completed. 2025-01-08T10:45:19,269 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=19}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34.' region-info for snapshot=emptySnaptb0-testExportWithTargetName 2025-01-08T10:45:19,269 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=18}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef.' region-info for snapshot=emptySnaptb0-testExportWithTargetName 2025-01-08T10:45:19,273 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=19}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:45:19,273 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=18}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:45:19,276 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=19}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T10:45:19,276 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=18}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T10:45:19,294 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741849_1025 (size=70) 2025-01-08T10:45:19,295 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741849_1025 (size=70) 2025-01-08T10:45:19,295 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741848_1024 (size=70) 2025-01-08T10:45:19,297 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741848_1024 (size=70) 2025-01-08T10:45:19,297 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=18}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef. 2025-01-08T10:45:19,297 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741849_1025 (size=70) 2025-01-08T10:45:19,297 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=19}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34. 2025-01-08T10:45:19,298 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741848_1024 (size=70) 2025-01-08T10:45:19,300 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=18}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=18 2025-01-08T10:45:19,300 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=19}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=19 2025-01-08T10:45:19,302 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=18 2025-01-08T10:45:19,302 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=19 2025-01-08T10:45:19,302 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportWithTargetName on region 6ac2ef231908ea1ef7af5e35d2ad7d34 2025-01-08T10:45:19,302 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportWithTargetName on region a23b54a820be126874ca84971483afef 2025-01-08T10:45:19,303 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=18, ppid=17, state=RUNNABLE; SnapshotRegionProcedure a23b54a820be126874ca84971483afef 2025-01-08T10:45:19,303 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=19, ppid=17, state=RUNNABLE; SnapshotRegionProcedure 6ac2ef231908ea1ef7af5e35d2ad7d34 2025-01-08T10:45:19,307 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=18, ppid=17, state=SUCCESS; SnapshotRegionProcedure a23b54a820be126874ca84971483afef in 203 msec 2025-01-08T10:45:19,310 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=19, resume processing ppid=17 2025-01-08T10:45:19,310 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=19, ppid=17, state=SUCCESS; SnapshotRegionProcedure 6ac2ef231908ea1ef7af5e35d2ad7d34 in 203 msec 2025-01-08T10:45:19,310 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T10:45:19,312 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T10:45:19,315 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T10:45:19,315 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T10:45:19,316 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:19,316 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(328): No files under family: cf 2025-01-08T10:45:19,327 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741850_1026 (size=62) 2025-01-08T10:45:19,327 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741850_1026 (size=62) 2025-01-08T10:45:19,328 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741850_1026 (size=62) 2025-01-08T10:45:19,331 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T10:45:19,331 DEBUG [PEWorker-4 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for emptySnaptb0-testExportWithTargetName 2025-01-08T10:45:19,334 DEBUG [PEWorker-4 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/emptySnaptb0-testExportWithTargetName 2025-01-08T10:45:19,357 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741851_1027 (size=649) 2025-01-08T10:45:19,357 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741851_1027 (size=649) 2025-01-08T10:45:19,357 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741851_1027 (size=649) 2025-01-08T10:45:19,362 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T10:45:19,370 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=17 2025-01-08T10:45:19,374 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T10:45:19,375 DEBUG [PEWorker-4 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/emptySnaptb0-testExportWithTargetName to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/emptySnaptb0-testExportWithTargetName 2025-01-08T10:45:19,378 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T10:45:19,378 DEBUG [PEWorker-4 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 }, snapshot procedure id = 17 2025-01-08T10:45:19,381 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=17, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } in 319 msec 2025-01-08T10:45:19,672 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=17 2025-01-08T10:45:19,673 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportWithTargetName, procId: 17 completed 2025-01-08T10:45:19,685 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=45003 {}] regionserver.HRegion(8254): writing data to region testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T10:45:19,687 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36283 {}] regionserver.HRegion(8254): writing data to region testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T10:45:19,696 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportWithTargetName 2025-01-08T10:45:19,696 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef. 2025-01-08T10:45:19,697 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T10:45:19,719 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.2 snapshot request for:{ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } 2025-01-08T10:45:19,719 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736333119719 (current time:1736333119719). 2025-01-08T10:45:19,719 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T10:45:19,719 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testExportWithTargetName VERSION not specified, setting to 2 2025-01-08T10:45:19,719 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T10:45:19,721 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x36c91cad to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@480f2430 2025-01-08T10:45:19,725 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@61fa1b1e, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:45:19,727 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:45:19,729 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:42488, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:45:19,731 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x36c91cad to 127.0.0.1:52367 2025-01-08T10:45:19,731 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:45:19,733 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x7668452b to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@4777d172 2025-01-08T10:45:19,737 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@76640069, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:45:19,745 DEBUG [hconnection-0x2e61e7f4-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:45:19,746 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:42494, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:45:19,751 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x7668452b to 127.0.0.1:52367 2025-01-08T10:45:19,752 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:45:19,752 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportWithTargetName], kv [jenkins: RWXCA] 2025-01-08T10:45:19,753 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T10:45:19,755 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=20, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } 2025-01-08T10:45:19,755 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 }, snapshot procedure id = 20 2025-01-08T10:45:19,757 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T10:45:19,759 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=20 2025-01-08T10:45:19,760 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T10:45:19,767 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T10:45:19,785 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741852_1028 (size=162) 2025-01-08T10:45:19,785 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741852_1028 (size=162) 2025-01-08T10:45:19,787 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741852_1028 (size=162) 2025-01-08T10:45:19,794 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T10:45:19,794 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=21, ppid=20, state=RUNNABLE; SnapshotRegionProcedure a23b54a820be126874ca84971483afef}, {pid=22, ppid=20, state=RUNNABLE; SnapshotRegionProcedure 6ac2ef231908ea1ef7af5e35d2ad7d34}] 2025-01-08T10:45:19,796 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=22, ppid=20, state=RUNNABLE; SnapshotRegionProcedure 6ac2ef231908ea1ef7af5e35d2ad7d34 2025-01-08T10:45:19,796 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=21, ppid=20, state=RUNNABLE; SnapshotRegionProcedure a23b54a820be126874ca84971483afef 2025-01-08T10:45:19,861 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=20 2025-01-08T10:45:19,948 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:45:19,948 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:19,949 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=45003 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=21 2025-01-08T10:45:19,949 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=36283 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=22 2025-01-08T10:45:19,949 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef. 2025-01-08T10:45:19,950 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34. 2025-01-08T10:45:19,950 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.HRegion(2837): Flushing a23b54a820be126874ca84971483afef 1/1 column families, dataSize=199 B heapSize=688 B 2025-01-08T10:45:19,950 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.HRegion(2837): Flushing 6ac2ef231908ea1ef7af5e35d2ad7d34 1/1 column families, dataSize=3.06 KB heapSize=6.86 KB 2025-01-08T10:45:20,030 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108258671aa178c43f695d4891b6979c651_6ac2ef231908ea1ef7af5e35d2ad7d34 is 71, key is 1646c0489d5ed2271fa126695f3c7bf0/cf:q/1736333119687/Put/seqid=0 2025-01-08T10:45:20,030 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e2025010821a6f6471a8143eea2b018224555e049_a23b54a820be126874ca84971483afef is 71, key is 08521e44d766d994dd36ae6ff4611e8f/cf:q/1736333119685/Put/seqid=0 2025-01-08T10:45:20,062 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741853_1029 (size=8172) 2025-01-08T10:45:20,064 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=20 2025-01-08T10:45:20,064 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741854_1030 (size=5102) 2025-01-08T10:45:20,064 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741853_1029 (size=8172) 2025-01-08T10:45:20,065 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741854_1030 (size=5102) 2025-01-08T10:45:20,066 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741853_1029 (size=8172) 2025-01-08T10:45:20,066 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:20,066 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741854_1030 (size=5102) 2025-01-08T10:45:20,066 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:20,157 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e2025010821a6f6471a8143eea2b018224555e049_a23b54a820be126874ca84971483afef to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/d41d8cd98f00b204e9800998ecf8427e2025010821a6f6471a8143eea2b018224555e049_a23b54a820be126874ca84971483afef 2025-01-08T10:45:20,157 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108258671aa178c43f695d4891b6979c651_6ac2ef231908ea1ef7af5e35d2ad7d34 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/c4ca4238a0b923820dcc509a6f75849b20250108258671aa178c43f695d4891b6979c651_6ac2ef231908ea1ef7af5e35d2ad7d34 2025-01-08T10:45:20,159 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/6ac2ef231908ea1ef7af5e35d2ad7d34/.tmp/cf/2965b77e3bb04cb2b40587934886775a, store: [table=testtb-testExportWithTargetName family=cf region=6ac2ef231908ea1ef7af5e35d2ad7d34] 2025-01-08T10:45:20,160 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/a23b54a820be126874ca84971483afef/.tmp/cf/a5e8a48b3e694b3581d381aa790c4ef9, store: [table=testtb-testExportWithTargetName family=cf region=a23b54a820be126874ca84971483afef] 2025-01-08T10:45:20,185 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/a23b54a820be126874ca84971483afef/.tmp/cf/a5e8a48b3e694b3581d381aa790c4ef9 is 208, key is 0f156c24f1a9f73d72cf743c920c051a1/cf:q/1736333119685/Put/seqid=0 2025-01-08T10:45:20,185 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/6ac2ef231908ea1ef7af5e35d2ad7d34/.tmp/cf/2965b77e3bb04cb2b40587934886775a is 208, key is 10c1aaef0ee73bd8b3af45b694eb180ec/cf:q/1736333119687/Put/seqid=0 2025-01-08T10:45:20,199 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741855_1031 (size=5914) 2025-01-08T10:45:20,200 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741855_1031 (size=5914) 2025-01-08T10:45:20,200 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741855_1031 (size=5914) 2025-01-08T10:45:20,201 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=199, hasBloomFilter=true, into tmp file hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/a23b54a820be126874ca84971483afef/.tmp/cf/a5e8a48b3e694b3581d381aa790c4ef9 2025-01-08T10:45:20,201 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741856_1032 (size=14951) 2025-01-08T10:45:20,202 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741856_1032 (size=14951) 2025-01-08T10:45:20,203 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741856_1032 (size=14951) 2025-01-08T10:45:20,207 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=3.1 K, hasBloomFilter=true, into tmp file hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/6ac2ef231908ea1ef7af5e35d2ad7d34/.tmp/cf/2965b77e3bb04cb2b40587934886775a 2025-01-08T10:45:20,217 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/a23b54a820be126874ca84971483afef/.tmp/cf/a5e8a48b3e694b3581d381aa790c4ef9 as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/a23b54a820be126874ca84971483afef/cf/a5e8a48b3e694b3581d381aa790c4ef9 2025-01-08T10:45:20,218 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/6ac2ef231908ea1ef7af5e35d2ad7d34/.tmp/cf/2965b77e3bb04cb2b40587934886775a as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/6ac2ef231908ea1ef7af5e35d2ad7d34/cf/2965b77e3bb04cb2b40587934886775a 2025-01-08T10:45:20,228 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/6ac2ef231908ea1ef7af5e35d2ad7d34/cf/2965b77e3bb04cb2b40587934886775a, entries=47, sequenceid=6, filesize=14.6 K 2025-01-08T10:45:20,228 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/a23b54a820be126874ca84971483afef/cf/a5e8a48b3e694b3581d381aa790c4ef9, entries=3, sequenceid=6, filesize=5.8 K 2025-01-08T10:45:20,232 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.HRegion(3040): Finished flush of dataSize ~199 B/199, heapSize ~672 B/672, currentSize=0 B/0 for a23b54a820be126874ca84971483afef in 282ms, sequenceid=6, compaction requested=false 2025-01-08T10:45:20,232 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportWithTargetName' 2025-01-08T10:45:20,232 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.HRegion(3040): Finished flush of dataSize ~3.06 KB/3137, heapSize ~6.84 KB/7008, currentSize=0 B/0 for 6ac2ef231908ea1ef7af5e35d2ad7d34 in 282ms, sequenceid=6, compaction requested=false 2025-01-08T10:45:20,232 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportWithTargetName' 2025-01-08T10:45:20,233 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.HRegion(2538): Flush status journal for 6ac2ef231908ea1ef7af5e35d2ad7d34: 2025-01-08T10:45:20,233 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.HRegion(2538): Flush status journal for a23b54a820be126874ca84971483afef: 2025-01-08T10:45:20,233 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef. for snaptb0-testExportWithTargetName completed. 2025-01-08T10:45:20,233 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34. for snaptb0-testExportWithTargetName completed. 2025-01-08T10:45:20,234 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef.' region-info for snapshot=snaptb0-testExportWithTargetName 2025-01-08T10:45:20,234 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34.' region-info for snapshot=snaptb0-testExportWithTargetName 2025-01-08T10:45:20,234 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:45:20,234 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:45:20,234 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/6ac2ef231908ea1ef7af5e35d2ad7d34/cf/2965b77e3bb04cb2b40587934886775a] hfiles 2025-01-08T10:45:20,234 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/a23b54a820be126874ca84971483afef/cf/a5e8a48b3e694b3581d381aa790c4ef9] hfiles 2025-01-08T10:45:20,234 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/6ac2ef231908ea1ef7af5e35d2ad7d34/cf/2965b77e3bb04cb2b40587934886775a for snapshot=snaptb0-testExportWithTargetName 2025-01-08T10:45:20,234 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/a23b54a820be126874ca84971483afef/cf/a5e8a48b3e694b3581d381aa790c4ef9 for snapshot=snaptb0-testExportWithTargetName 2025-01-08T10:45:20,249 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741857_1033 (size=109) 2025-01-08T10:45:20,250 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741858_1034 (size=109) 2025-01-08T10:45:20,251 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741857_1033 (size=109) 2025-01-08T10:45:20,251 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34. 2025-01-08T10:45:20,251 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=22 2025-01-08T10:45:20,252 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=22 2025-01-08T10:45:20,252 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportWithTargetName on region 6ac2ef231908ea1ef7af5e35d2ad7d34 2025-01-08T10:45:20,252 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=22, ppid=20, state=RUNNABLE; SnapshotRegionProcedure 6ac2ef231908ea1ef7af5e35d2ad7d34 2025-01-08T10:45:20,255 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741857_1033 (size=109) 2025-01-08T10:45:20,255 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef. 2025-01-08T10:45:20,255 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741858_1034 (size=109) 2025-01-08T10:45:20,255 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=21 2025-01-08T10:45:20,255 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741858_1034 (size=109) 2025-01-08T10:45:20,256 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=21 2025-01-08T10:45:20,256 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportWithTargetName on region a23b54a820be126874ca84971483afef 2025-01-08T10:45:20,256 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=21, ppid=20, state=RUNNABLE; SnapshotRegionProcedure a23b54a820be126874ca84971483afef 2025-01-08T10:45:20,257 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=22, ppid=20, state=SUCCESS; SnapshotRegionProcedure 6ac2ef231908ea1ef7af5e35d2ad7d34 in 460 msec 2025-01-08T10:45:20,261 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=21, resume processing ppid=20 2025-01-08T10:45:20,261 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T10:45:20,261 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=21, ppid=20, state=SUCCESS; SnapshotRegionProcedure a23b54a820be126874ca84971483afef in 463 msec 2025-01-08T10:45:20,263 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T10:45:20,265 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T10:45:20,265 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T10:45:20,265 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:20,269 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/c4ca4238a0b923820dcc509a6f75849b20250108258671aa178c43f695d4891b6979c651_6ac2ef231908ea1ef7af5e35d2ad7d34, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/d41d8cd98f00b204e9800998ecf8427e2025010821a6f6471a8143eea2b018224555e049_a23b54a820be126874ca84971483afef] hfiles 2025-01-08T10:45:20,269 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/c4ca4238a0b923820dcc509a6f75849b20250108258671aa178c43f695d4891b6979c651_6ac2ef231908ea1ef7af5e35d2ad7d34 2025-01-08T10:45:20,269 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/d41d8cd98f00b204e9800998ecf8427e2025010821a6f6471a8143eea2b018224555e049_a23b54a820be126874ca84971483afef 2025-01-08T10:45:20,286 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741859_1035 (size=293) 2025-01-08T10:45:20,287 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741859_1035 (size=293) 2025-01-08T10:45:20,287 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741859_1035 (size=293) 2025-01-08T10:45:20,289 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T10:45:20,289 DEBUG [PEWorker-2 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testExportWithTargetName 2025-01-08T10:45:20,291 DEBUG [PEWorker-2 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/snaptb0-testExportWithTargetName 2025-01-08T10:45:20,312 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741860_1036 (size=959) 2025-01-08T10:45:20,312 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741860_1036 (size=959) 2025-01-08T10:45:20,313 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741860_1036 (size=959) 2025-01-08T10:45:20,318 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T10:45:20,330 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T10:45:20,331 DEBUG [PEWorker-2 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/snaptb0-testExportWithTargetName to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportWithTargetName 2025-01-08T10:45:20,334 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T10:45:20,334 DEBUG [PEWorker-2 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 }, snapshot procedure id = 20 2025-01-08T10:45:20,337 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=20, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } in 581 msec 2025-01-08T10:45:20,366 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=20 2025-01-08T10:45:20,366 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportWithTargetName, procId: 20 completed 2025-01-08T10:45:20,367 INFO [Time-limited test {}] snapshot.TestExportSnapshot(468): HDFS export destination path: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333120367 2025-01-08T10:45:20,367 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=hdfs://localhost:37495, tgtDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333120367, rawTgtDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333120367, srcFsUri=hdfs://localhost:37495, srcDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:45:20,417 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:37495, inputRoot=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:45:20,417 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-832679797_22, ugi=jenkins (auth:SIMPLE)]], outputRoot=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333120367, skipTmp=false, initialOutputSnapshotDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333120367/.hbase-snapshot/.tmp/testExportWithTargetName 2025-01-08T10:45:20,423 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T10:45:20,443 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportWithTargetName to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333120367/.hbase-snapshot/.tmp/testExportWithTargetName 2025-01-08T10:45:20,477 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741861_1037 (size=162) 2025-01-08T10:45:20,477 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741861_1037 (size=162) 2025-01-08T10:45:20,478 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741861_1037 (size=162) 2025-01-08T10:45:20,499 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741862_1038 (size=959) 2025-01-08T10:45:20,500 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741862_1038 (size=959) 2025-01-08T10:45:20,501 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741862_1038 (size=959) 2025-01-08T10:45:20,522 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741863_1039 (size=154) 2025-01-08T10:45:20,522 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741863_1039 (size=154) 2025-01-08T10:45:20,523 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741863_1039 (size=154) 2025-01-08T10:45:20,838 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-16693397335135134725.jar 2025-01-08T10:45:20,839 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T10:45:20,839 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T10:45:20,840 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T10:45:21,865 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-12870750722149265941.jar 2025-01-08T10:45:21,866 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T10:45:21,866 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T10:45:21,937 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-2573554516258630434.jar 2025-01-08T10:45:21,938 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T10:45:21,938 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T10:45:21,938 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T10:45:21,939 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T10:45:21,939 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T10:45:21,939 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T10:45:21,940 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T10:45:21,940 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T10:45:21,940 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T10:45:21,941 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T10:45:21,941 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T10:45:21,942 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T10:45:21,942 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T10:45:21,942 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T10:45:21,943 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T10:45:21,943 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T10:45:21,943 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T10:45:21,944 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T10:45:21,946 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:45:21,946 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:45:21,947 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:45:21,947 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:45:21,947 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:45:21,948 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:45:21,948 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:45:22,100 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741864_1040 (size=127628) 2025-01-08T10:45:22,100 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741864_1040 (size=127628) 2025-01-08T10:45:22,101 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741864_1040 (size=127628) 2025-01-08T10:45:22,124 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741865_1041 (size=2172137) 2025-01-08T10:45:22,125 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741865_1041 (size=2172137) 2025-01-08T10:45:22,125 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741865_1041 (size=2172137) 2025-01-08T10:45:22,162 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741866_1042 (size=213228) 2025-01-08T10:45:22,165 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741866_1042 (size=213228) 2025-01-08T10:45:22,167 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741866_1042 (size=213228) 2025-01-08T10:45:22,210 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741867_1043 (size=1877034) 2025-01-08T10:45:22,210 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741867_1043 (size=1877034) 2025-01-08T10:45:22,211 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741867_1043 (size=1877034) 2025-01-08T10:45:22,227 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741868_1044 (size=533455) 2025-01-08T10:45:22,227 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741868_1044 (size=533455) 2025-01-08T10:45:22,228 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741868_1044 (size=533455) 2025-01-08T10:45:22,243 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741869_1045 (size=912094) 2025-01-08T10:45:22,243 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741869_1045 (size=912094) 2025-01-08T10:45:22,244 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741869_1045 (size=912094) 2025-01-08T10:45:22,285 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741870_1046 (size=7280644) 2025-01-08T10:45:22,287 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741870_1046 (size=7280644) 2025-01-08T10:45:22,287 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741870_1046 (size=7280644) 2025-01-08T10:45:22,317 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741871_1047 (size=4188619) 2025-01-08T10:45:22,317 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741871_1047 (size=4188619) 2025-01-08T10:45:22,318 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741871_1047 (size=4188619) 2025-01-08T10:45:22,331 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741872_1048 (size=20406) 2025-01-08T10:45:22,331 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741872_1048 (size=20406) 2025-01-08T10:45:22,331 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741872_1048 (size=20406) 2025-01-08T10:45:22,344 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741873_1049 (size=75495) 2025-01-08T10:45:22,344 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741873_1049 (size=75495) 2025-01-08T10:45:22,344 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741873_1049 (size=75495) 2025-01-08T10:45:22,359 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741874_1050 (size=45609) 2025-01-08T10:45:22,360 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741874_1050 (size=45609) 2025-01-08T10:45:22,360 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741874_1050 (size=45609) 2025-01-08T10:45:22,376 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741875_1051 (size=110084) 2025-01-08T10:45:22,377 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741875_1051 (size=110084) 2025-01-08T10:45:22,377 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741875_1051 (size=110084) 2025-01-08T10:45:22,392 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741876_1052 (size=1323991) 2025-01-08T10:45:22,392 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741876_1052 (size=1323991) 2025-01-08T10:45:22,393 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741876_1052 (size=1323991) 2025-01-08T10:45:22,405 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741877_1053 (size=23076) 2025-01-08T10:45:22,405 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741877_1053 (size=23076) 2025-01-08T10:45:22,406 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741877_1053 (size=23076) 2025-01-08T10:45:22,416 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741878_1054 (size=126803) 2025-01-08T10:45:22,416 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741878_1054 (size=126803) 2025-01-08T10:45:22,417 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741878_1054 (size=126803) 2025-01-08T10:45:22,434 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741879_1055 (size=322274) 2025-01-08T10:45:22,434 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741879_1055 (size=322274) 2025-01-08T10:45:22,434 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741879_1055 (size=322274) 2025-01-08T10:45:22,447 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741880_1056 (size=451756) 2025-01-08T10:45:22,448 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741880_1056 (size=451756) 2025-01-08T10:45:22,448 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741880_1056 (size=451756) 2025-01-08T10:45:22,465 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741881_1057 (size=1832290) 2025-01-08T10:45:22,466 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741881_1057 (size=1832290) 2025-01-08T10:45:22,467 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741881_1057 (size=1832290) 2025-01-08T10:45:22,477 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741882_1058 (size=30081) 2025-01-08T10:45:22,477 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741882_1058 (size=30081) 2025-01-08T10:45:22,478 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741882_1058 (size=30081) 2025-01-08T10:45:22,493 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741883_1059 (size=53616) 2025-01-08T10:45:22,493 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741883_1059 (size=53616) 2025-01-08T10:45:22,493 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741883_1059 (size=53616) 2025-01-08T10:45:22,509 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741884_1060 (size=29229) 2025-01-08T10:45:22,510 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741884_1060 (size=29229) 2025-01-08T10:45:22,511 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741884_1060 (size=29229) 2025-01-08T10:45:22,578 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741885_1061 (size=169089) 2025-01-08T10:45:22,578 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741885_1061 (size=169089) 2025-01-08T10:45:22,579 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741885_1061 (size=169089) 2025-01-08T10:45:22,625 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741886_1062 (size=5175431) 2025-01-08T10:45:22,625 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741886_1062 (size=5175431) 2025-01-08T10:45:22,626 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741886_1062 (size=5175431) 2025-01-08T10:45:22,642 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741887_1063 (size=136454) 2025-01-08T10:45:22,642 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741887_1063 (size=136454) 2025-01-08T10:45:22,643 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741887_1063 (size=136454) 2025-01-08T10:45:22,665 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741888_1064 (size=3317408) 2025-01-08T10:45:22,666 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741888_1064 (size=3317408) 2025-01-08T10:45:22,666 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741888_1064 (size=3317408) 2025-01-08T10:45:23,111 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741889_1065 (size=6350714) 2025-01-08T10:45:23,111 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741889_1065 (size=6350714) 2025-01-08T10:45:23,112 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741889_1065 (size=6350714) 2025-01-08T10:45:23,135 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741890_1066 (size=503880) 2025-01-08T10:45:23,136 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741890_1066 (size=503880) 2025-01-08T10:45:23,136 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741890_1066 (size=503880) 2025-01-08T10:45:23,176 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741891_1067 (size=4695811) 2025-01-08T10:45:23,176 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741891_1067 (size=4695811) 2025-01-08T10:45:23,177 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741891_1067 (size=4695811) 2025-01-08T10:45:23,182 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T10:45:23,191 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'snaptb0-testExportWithTargetName' hfile list 2025-01-08T10:45:23,201 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(751): export split=0 size=33.3 K 2025-01-08T10:45:23,231 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741892_1068 (size=722) 2025-01-08T10:45:23,237 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741892_1068 (size=722) 2025-01-08T10:45:23,238 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741892_1068 (size=722) 2025-01-08T10:45:23,253 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741893_1069 (size=15) 2025-01-08T10:45:23,254 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741893_1069 (size=15) 2025-01-08T10:45:23,257 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741893_1069 (size=15) 2025-01-08T10:45:23,455 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741894_1070 (size=304934) 2025-01-08T10:45:23,458 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741894_1070 (size=304934) 2025-01-08T10:45:23,459 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741894_1070 (size=304934) 2025-01-08T10:45:23,840 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T10:45:23,840 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T10:45:24,059 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0001_000001 (auth:SIMPLE) from 127.0.0.1:59902 2025-01-08T10:45:24,252 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T10:45:27,596 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportWithTargetName 2025-01-08T10:45:27,596 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportWithTargetName Metrics about Tables on a single HBase RegionServer 2025-01-08T10:45:33,908 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0001_000001 (auth:SIMPLE) from 127.0.0.1:49898 2025-01-08T10:45:34,209 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741895_1071 (size=350608) 2025-01-08T10:45:34,209 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741895_1071 (size=350608) 2025-01-08T10:45:34,209 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741895_1071 (size=350608) 2025-01-08T10:45:35,713 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T10:45:36,225 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0001_000001 (auth:SIMPLE) from 127.0.0.1:48266 2025-01-08T10:45:40,958 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741896_1072 (size=14951) 2025-01-08T10:45:40,958 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741896_1072 (size=14951) 2025-01-08T10:45:40,958 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741896_1072 (size=14951) 2025-01-08T10:45:41,031 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741897_1073 (size=8172) 2025-01-08T10:45:41,032 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741897_1073 (size=8172) 2025-01-08T10:45:41,032 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741897_1073 (size=8172) 2025-01-08T10:45:41,082 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741898_1074 (size=5914) 2025-01-08T10:45:41,083 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741898_1074 (size=5914) 2025-01-08T10:45:41,083 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741898_1074 (size=5914) 2025-01-08T10:45:41,137 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741899_1075 (size=5102) 2025-01-08T10:45:41,139 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741899_1075 (size=5102) 2025-01-08T10:45:41,139 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741899_1075 (size=5102) 2025-01-08T10:45:41,250 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741900_1076 (size=17461) 2025-01-08T10:45:41,250 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741900_1076 (size=17461) 2025-01-08T10:45:41,251 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741900_1076 (size=17461) 2025-01-08T10:45:41,268 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741901_1077 (size=464) 2025-01-08T10:45:41,269 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741901_1077 (size=464) 2025-01-08T10:45:41,269 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741901_1077 (size=464) 2025-01-08T10:45:41,333 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741902_1078 (size=17461) 2025-01-08T10:45:41,333 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741902_1078 (size=17461) 2025-01-08T10:45:41,334 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741902_1078 (size=17461) 2025-01-08T10:45:41,358 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741903_1079 (size=350608) 2025-01-08T10:45:41,360 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741903_1079 (size=350608) 2025-01-08T10:45:41,360 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741903_1079 (size=350608) 2025-01-08T10:45:41,380 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0001_000001 (auth:SIMPLE) from 127.0.0.1:48268 2025-01-08T10:45:41,398 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_1/usercache/jenkins/appcache/application_1736333116683_0001/container_1736333116683_0001_01_000002/launch_container.sh] 2025-01-08T10:45:41,406 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_1/usercache/jenkins/appcache/application_1736333116683_0001/container_1736333116683_0001_01_000002/container_tokens] 2025-01-08T10:45:41,407 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_1/usercache/jenkins/appcache/application_1736333116683_0001/container_1736333116683_0001_01_000002/sysfs] 2025-01-08T10:45:41,524 DEBUG [BootstrapNodeManager {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2025-01-08T10:45:41,526 INFO [RS-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:60212, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=RegionServerStatusService 2025-01-08T10:45:42,522 DEBUG [BootstrapNodeManager {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2025-01-08T10:45:42,523 INFO [RS-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:60222, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.1 (auth:SIMPLE), service=RegionServerStatusService 2025-01-08T10:45:43,302 INFO [Time-limited test {}] snapshot.ExportSnapshot(1207): Finalize the Snapshot Export 2025-01-08T10:45:43,303 INFO [Time-limited test {}] snapshot.ExportSnapshot(1218): Verify the exported snapshot's expiration status and integrity. 2025-01-08T10:45:43,320 INFO [Time-limited test {}] snapshot.ExportSnapshot(1224): Export Completed: testExportWithTargetName 2025-01-08T10:45:43,320 INFO [Time-limited test {}] snapshot.TestExportSnapshot(362): Exported snapshot 2025-01-08T10:45:43,321 INFO [Time-limited test {}] snapshot.TestExportSnapshot(373): Verified filesystem state 2025-01-08T10:45:43,321 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-832679797_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportWithTargetName at hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportWithTargetName 2025-01-08T10:45:43,322 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportWithTargetName/.snapshotinfo 2025-01-08T10:45:43,322 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportWithTargetName/data.manifest 2025-01-08T10:45:43,322 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-832679797_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333120367/.hbase-snapshot/testExportWithTargetName at hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333120367/.hbase-snapshot/testExportWithTargetName 2025-01-08T10:45:43,323 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333120367/.hbase-snapshot/testExportWithTargetName/.snapshotinfo 2025-01-08T10:45:43,323 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333120367/.hbase-snapshot/testExportWithTargetName/data.manifest 2025-01-08T10:45:43,338 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testExportWithTargetName 2025-01-08T10:45:43,341 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.2 disable testtb-testExportWithTargetName 2025-01-08T10:45:43,348 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=23, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testExportWithTargetName 2025-01-08T10:45:43,353 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=23 2025-01-08T10:45:43,354 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithTargetName","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333143354"}]},"ts":"1736333143354"} 2025-01-08T10:45:43,356 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithTargetName, state=DISABLING in hbase:meta 2025-01-08T10:45:43,359 INFO [PEWorker-1 {}] procedure.DisableTableProcedure(284): Set testtb-testExportWithTargetName to state=DISABLING 2025-01-08T10:45:43,361 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=24, ppid=23, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testExportWithTargetName}] 2025-01-08T10:45:43,368 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=25, ppid=24, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=a23b54a820be126874ca84971483afef, UNASSIGN}, {pid=26, ppid=24, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=6ac2ef231908ea1ef7af5e35d2ad7d34, UNASSIGN}] 2025-01-08T10:45:43,370 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=26, ppid=24, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=6ac2ef231908ea1ef7af5e35d2ad7d34, UNASSIGN 2025-01-08T10:45:43,370 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=25, ppid=24, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=a23b54a820be126874ca84971483afef, UNASSIGN 2025-01-08T10:45:43,372 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=26 updating hbase:meta row=6ac2ef231908ea1ef7af5e35d2ad7d34, regionState=CLOSING, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:43,372 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=25 updating hbase:meta row=a23b54a820be126874ca84971483afef, regionState=CLOSING, regionLocation=67f42ee69b4e,45003,1736333108039 2025-01-08T10:45:43,374 DEBUG [PEWorker-5 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T10:45:43,375 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=27, ppid=26, state=RUNNABLE; CloseRegionProcedure 6ac2ef231908ea1ef7af5e35d2ad7d34, server=67f42ee69b4e,36283,1736333108166}] 2025-01-08T10:45:43,376 DEBUG [PEWorker-4 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T10:45:43,379 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=28, ppid=25, state=RUNNABLE; CloseRegionProcedure a23b54a820be126874ca84971483afef, server=67f42ee69b4e,45003,1736333108039}] 2025-01-08T10:45:43,455 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=23 2025-01-08T10:45:43,533 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:43,535 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] handler.UnassignRegionHandler(124): Close 6ac2ef231908ea1ef7af5e35d2ad7d34 2025-01-08T10:45:43,535 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T10:45:43,536 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:45:43,536 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] regionserver.HRegion(1681): Closing 6ac2ef231908ea1ef7af5e35d2ad7d34, disabling compactions & flushes 2025-01-08T10:45:43,536 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] regionserver.HRegion(1703): Closing region testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34. 2025-01-08T10:45:43,537 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34. 2025-01-08T10:45:43,537 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34. after waiting 0 ms 2025-01-08T10:45:43,537 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34. 2025-01-08T10:45:43,537 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] handler.UnassignRegionHandler(124): Close a23b54a820be126874ca84971483afef 2025-01-08T10:45:43,537 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T10:45:43,537 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] regionserver.HRegion(1681): Closing a23b54a820be126874ca84971483afef, disabling compactions & flushes 2025-01-08T10:45:43,537 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] regionserver.HRegion(1703): Closing region testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef. 2025-01-08T10:45:43,538 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef. 2025-01-08T10:45:43,538 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef. after waiting 0 ms 2025-01-08T10:45:43,538 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef. 2025-01-08T10:45:43,548 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/a23b54a820be126874ca84971483afef/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T10:45:43,550 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/6ac2ef231908ea1ef7af5e35d2ad7d34/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T10:45:43,552 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:45:43,552 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:45:43,552 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] regionserver.HRegion(1922): Closed testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34. 2025-01-08T10:45:43,552 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] regionserver.HRegion(1922): Closed testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef. 2025-01-08T10:45:43,553 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] regionserver.HRegion(1635): Region close journal for 6ac2ef231908ea1ef7af5e35d2ad7d34: 2025-01-08T10:45:43,553 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] regionserver.HRegion(1635): Region close journal for a23b54a820be126874ca84971483afef: 2025-01-08T10:45:43,558 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] handler.UnassignRegionHandler(170): Closed a23b54a820be126874ca84971483afef 2025-01-08T10:45:43,558 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=25 updating hbase:meta row=a23b54a820be126874ca84971483afef, regionState=CLOSED 2025-01-08T10:45:43,575 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] handler.UnassignRegionHandler(170): Closed 6ac2ef231908ea1ef7af5e35d2ad7d34 2025-01-08T10:45:43,576 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=26 updating hbase:meta row=6ac2ef231908ea1ef7af5e35d2ad7d34, regionState=CLOSED 2025-01-08T10:45:43,577 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=28, resume processing ppid=25 2025-01-08T10:45:43,577 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=28, ppid=25, state=SUCCESS; CloseRegionProcedure a23b54a820be126874ca84971483afef, server=67f42ee69b4e,45003,1736333108039 in 185 msec 2025-01-08T10:45:43,579 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=25, ppid=24, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=a23b54a820be126874ca84971483afef, UNASSIGN in 209 msec 2025-01-08T10:45:43,581 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=27, resume processing ppid=26 2025-01-08T10:45:43,581 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=27, ppid=26, state=SUCCESS; CloseRegionProcedure 6ac2ef231908ea1ef7af5e35d2ad7d34, server=67f42ee69b4e,36283,1736333108166 in 203 msec 2025-01-08T10:45:43,584 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=26, resume processing ppid=24 2025-01-08T10:45:43,584 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=26, ppid=24, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=6ac2ef231908ea1ef7af5e35d2ad7d34, UNASSIGN in 213 msec 2025-01-08T10:45:43,588 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=24, resume processing ppid=23 2025-01-08T10:45:43,588 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=24, ppid=23, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testExportWithTargetName in 224 msec 2025-01-08T10:45:43,590 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithTargetName","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333143589"}]},"ts":"1736333143589"} 2025-01-08T10:45:43,592 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithTargetName, state=DISABLED in hbase:meta 2025-01-08T10:45:43,594 INFO [PEWorker-3 {}] procedure.DisableTableProcedure(296): Set testtb-testExportWithTargetName to state=DISABLED 2025-01-08T10:45:43,597 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=23, state=SUCCESS; DisableTableProcedure table=testtb-testExportWithTargetName in 251 msec 2025-01-08T10:45:43,656 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=23 2025-01-08T10:45:43,656 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testExportWithTargetName, procId: 23 completed 2025-01-08T10:45:43,660 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.2 delete testtb-testExportWithTargetName 2025-01-08T10:45:43,664 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=29, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testExportWithTargetName 2025-01-08T10:45:43,666 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=29, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportWithTargetName 2025-01-08T10:45:43,667 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testExportWithTargetName 2025-01-08T10:45:43,668 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=29, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testExportWithTargetName 2025-01-08T10:45:43,671 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36283 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testExportWithTargetName 2025-01-08T10:45:43,674 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithTargetName 2025-01-08T10:45:43,674 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithTargetName 2025-01-08T10:45:43,674 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithTargetName 2025-01-08T10:45:43,674 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithTargetName 2025-01-08T10:45:43,676 DEBUG [HFileArchiver-1 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/a23b54a820be126874ca84971483afef 2025-01-08T10:45:43,676 DEBUG [HFileArchiver-2 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/6ac2ef231908ea1ef7af5e35d2ad7d34 2025-01-08T10:45:43,676 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:45:43,676 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:45:43,676 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:45:43,676 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithTargetName with data null 2025-01-08T10:45:43,676 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithTargetName with data null 2025-01-08T10:45:43,676 INFO [zk-permission-watcher-pool-0 {}] access.AuthManager(135): Skipping permission cache refresh because writable data is empty 2025-01-08T10:45:43,676 INFO [zk-permission-watcher-pool-0 {}] access.AuthManager(135): Skipping permission cache refresh because writable data is empty 2025-01-08T10:45:43,677 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithTargetName with data null 2025-01-08T10:45:43,677 INFO [zk-permission-watcher-pool-0 {}] access.AuthManager(135): Skipping permission cache refresh because writable data is empty 2025-01-08T10:45:43,677 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:45:43,677 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithTargetName with data null 2025-01-08T10:45:43,677 INFO [zk-permission-watcher-pool-0 {}] access.AuthManager(135): Skipping permission cache refresh because writable data is empty 2025-01-08T10:45:43,680 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=29 2025-01-08T10:45:43,681 DEBUG [HFileArchiver-1 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/a23b54a820be126874ca84971483afef/cf, FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/a23b54a820be126874ca84971483afef/recovered.edits] 2025-01-08T10:45:43,681 DEBUG [HFileArchiver-2 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/6ac2ef231908ea1ef7af5e35d2ad7d34/cf, FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/6ac2ef231908ea1ef7af5e35d2ad7d34/recovered.edits] 2025-01-08T10:45:43,689 DEBUG [HFileArchiver-2 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/6ac2ef231908ea1ef7af5e35d2ad7d34/cf/2965b77e3bb04cb2b40587934886775a to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportWithTargetName/6ac2ef231908ea1ef7af5e35d2ad7d34/cf/2965b77e3bb04cb2b40587934886775a 2025-01-08T10:45:43,689 DEBUG [HFileArchiver-1 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/a23b54a820be126874ca84971483afef/cf/a5e8a48b3e694b3581d381aa790c4ef9 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportWithTargetName/a23b54a820be126874ca84971483afef/cf/a5e8a48b3e694b3581d381aa790c4ef9 2025-01-08T10:45:43,693 DEBUG [HFileArchiver-1 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/a23b54a820be126874ca84971483afef/recovered.edits/9.seqid to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportWithTargetName/a23b54a820be126874ca84971483afef/recovered.edits/9.seqid 2025-01-08T10:45:43,693 DEBUG [HFileArchiver-2 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/6ac2ef231908ea1ef7af5e35d2ad7d34/recovered.edits/9.seqid to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportWithTargetName/6ac2ef231908ea1ef7af5e35d2ad7d34/recovered.edits/9.seqid 2025-01-08T10:45:43,693 DEBUG [HFileArchiver-1 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/a23b54a820be126874ca84971483afef 2025-01-08T10:45:43,694 DEBUG [HFileArchiver-2 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithTargetName/6ac2ef231908ea1ef7af5e35d2ad7d34 2025-01-08T10:45:43,694 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(313): Archived testtb-testExportWithTargetName regions 2025-01-08T10:45:43,694 DEBUG [PEWorker-4 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71 2025-01-08T10:45:43,695 DEBUG [PEWorker-4 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf] 2025-01-08T10:45:43,700 DEBUG [PEWorker-4 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/c4ca4238a0b923820dcc509a6f75849b20250108258671aa178c43f695d4891b6979c651_6ac2ef231908ea1ef7af5e35d2ad7d34 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/c4ca4238a0b923820dcc509a6f75849b20250108258671aa178c43f695d4891b6979c651_6ac2ef231908ea1ef7af5e35d2ad7d34 2025-01-08T10:45:43,701 DEBUG [PEWorker-4 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/d41d8cd98f00b204e9800998ecf8427e2025010821a6f6471a8143eea2b018224555e049_a23b54a820be126874ca84971483afef to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/d41d8cd98f00b204e9800998ecf8427e2025010821a6f6471a8143eea2b018224555e049_a23b54a820be126874ca84971483afef 2025-01-08T10:45:43,702 DEBUG [PEWorker-4 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71 2025-01-08T10:45:43,705 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=29, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testExportWithTargetName 2025-01-08T10:45:43,709 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=36283 {}] util.ReflectedFunctionCache(97): Populated cache for org.apache.hadoop.hbase.filter.KeyOnlyFilter in 0ms 2025-01-08T10:45:43,712 WARN [PEWorker-4 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testtb-testExportWithTargetName from hbase:meta 2025-01-08T10:45:43,715 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testExportWithTargetName' descriptor. 2025-01-08T10:45:43,716 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=29, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testExportWithTargetName 2025-01-08T10:45:43,716 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testExportWithTargetName' from region states. 2025-01-08T10:45:43,717 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736333143717"}]},"ts":"9223372036854775807"} 2025-01-08T10:45:43,717 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736333143717"}]},"ts":"9223372036854775807"} 2025-01-08T10:45:43,719 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T10:45:43,719 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => a23b54a820be126874ca84971483afef, NAME => 'testtb-testExportWithTargetName,,1736333118340.a23b54a820be126874ca84971483afef.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => 6ac2ef231908ea1ef7af5e35d2ad7d34, NAME => 'testtb-testExportWithTargetName,1,1736333118340.6ac2ef231908ea1ef7af5e35d2ad7d34.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T10:45:43,719 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testExportWithTargetName' as deleted. 2025-01-08T10:45:43,719 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportWithTargetName","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736333143719"}]},"ts":"9223372036854775807"} 2025-01-08T10:45:43,721 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testExportWithTargetName state from META 2025-01-08T10:45:43,724 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(133): Finished pid=29, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportWithTargetName 2025-01-08T10:45:43,726 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=29, state=SUCCESS; DeleteTableProcedure table=testtb-testExportWithTargetName in 64 msec 2025-01-08T10:45:43,781 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=29 2025-01-08T10:45:43,782 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testExportWithTargetName, procId: 29 completed 2025-01-08T10:45:43,800 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.2 delete name: "emptySnaptb0-testExportWithTargetName" 2025-01-08T10:45:43,803 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotManager(380): Deleting snapshot: emptySnaptb0-testExportWithTargetName 2025-01-08T10:45:43,806 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.2 delete name: "snaptb0-testExportWithTargetName" 2025-01-08T10:45:43,809 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testExportWithTargetName 2025-01-08T10:45:43,847 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: snapshot.TestMobSecureExportSnapshot#testExportWithTargetName Thread=771 (was 718) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-7 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-6 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (175901922) connection to localhost/127.0.0.1:45235 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: zk-permission-watcher-pool-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: zk-permission-watcher-pool-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Async disk worker #0 for volume /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data4 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ContainersLauncher #0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:401) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:34415 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-6-2 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:62) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:883) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Async disk worker #0 for volume /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data5 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-3-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Async disk worker #0 for volume /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:42701 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DeletionService #3 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Thread-1331 java.base@17.0.11/java.io.FileInputStream.readBytes(Native Method) java.base@17.0.11/java.io.FileInputStream.read(FileInputStream.java:276) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:282) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) java.base@17.0.11/sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:281) java.base@17.0.11/sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:324) java.base@17.0.11/sun.nio.cs.StreamDecoder.read(StreamDecoder.java:189) java.base@17.0.11/java.io.InputStreamReader.read(InputStreamReader.java:177) java.base@17.0.11/java.io.BufferedReader.fill(BufferedReader.java:162) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:329) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:396) app//org.apache.hadoop.util.Shell$1.run(Shell.java:1025) Potentially hanging thread: IPC Client (175901922) connection to localhost/127.0.0.1:42701 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-832679797_22 at /127.0.0.1:33212 [Waiting for operation #3] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-549257842_1 at /127.0.0.1:36558 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-5 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-3-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: zk-permission-watcher-pool-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-832679797_22 at /127.0.0.1:36576 [Waiting for operation #6] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: htable-pool-1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:401) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-6-3 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:62) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:883) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DeletionService #1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DeletionService #0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Container metrics unregistration java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-549257842_1 at /127.0.0.1:33186 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ContainersLauncher #0 java.base@17.0.11/java.io.FileInputStream.readBytes(Native Method) java.base@17.0.11/java.io.FileInputStream.read(FileInputStream.java:276) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:282) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) java.base@17.0.11/sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:281) java.base@17.0.11/sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:324) java.base@17.0.11/sun.nio.cs.StreamDecoder.read(StreamDecoder.java:189) java.base@17.0.11/java.io.InputStreamReader.read(InputStreamReader.java:177) java.base@17.0.11/java.io.BufferedReader.fill(BufferedReader.java:162) java.base@17.0.11/java.io.BufferedReader.read1(BufferedReader.java:213) java.base@17.0.11/java.io.BufferedReader.read(BufferedReader.java:287) app//org.apache.hadoop.util.Shell$ShellCommandExecutor.parseExecResult(Shell.java:1295) app//org.apache.hadoop.util.Shell.runCommand(Shell.java:1054) app//org.apache.hadoop.util.Shell.run(Shell.java:959) app//org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:1282) app//org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:349) app//org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.launchContainer(ContainerLaunch.java:600) app//org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:388) app//org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:105) java.base@17.0.11/java.util.concurrent.FutureTask.run(FutureTask.java:264) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: htable-pool-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:401) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DeletionService #0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DeletionService #2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Async disk worker #0 for volume /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:45235 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Async disk worker #0 for volume /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data3 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: process reaper (pid 5911) java.base@17.0.11/java.lang.ProcessHandleImpl.waitForProcessExit0(Native Method) java.base@17.0.11/java.lang.ProcessHandleImpl$1.run(ProcessHandleImpl.java:150) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Async disk worker #0 for volume /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data6 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (175901922) connection to localhost/127.0.0.1:34415 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: zk-permission-watcher-pool-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-832679797_22 at /127.0.0.1:59658 [Waiting for operation #3] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-8 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) - Thread LEAK? -, OpenFileDescriptor=813 (was 781) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=802 (was 735) - SystemLoadAverage LEAK? -, ProcessCount=20 (was 19) - ProcessCount LEAK? -, AvailableMemoryMB=1212 (was 982) - AvailableMemoryMB LEAK? - 2025-01-08T10:45:43,848 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=771 is superior to 500 2025-01-08T10:45:43,865 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: snapshot.TestMobSecureExportSnapshot#testExportWithResetTtl Thread=771, OpenFileDescriptor=813, MaxFileDescriptor=1048576, SystemLoadAverage=802, ProcessCount=20, AvailableMemoryMB=1210 2025-01-08T10:45:43,865 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=771 is superior to 500 2025-01-08T10:45:43,868 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.2 create 'testtb-testExportWithResetTtl', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T10:45:43,869 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=30, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testExportWithResetTtl 2025-01-08T10:45:43,870 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=30, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportWithResetTtl execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T10:45:43,870 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "testtb-testExportWithResetTtl" procId is: 30 2025-01-08T10:45:43,871 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=30, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testExportWithResetTtl execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T10:45:43,872 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=30 2025-01-08T10:45:43,883 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741904_1080 (size=440) 2025-01-08T10:45:43,884 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741904_1080 (size=440) 2025-01-08T10:45:43,884 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741904_1080 (size=440) 2025-01-08T10:45:43,886 INFO [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => af16492a38d4fc130200d4b3697ff170, NAME => 'testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testtb-testExportWithResetTtl', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:45:43,887 INFO [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => 45518da4f8eaa6bc80e04f2fcdc6fc93, NAME => 'testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testtb-testExportWithResetTtl', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:45:43,895 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741905_1081 (size=65) 2025-01-08T10:45:43,895 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741905_1081 (size=65) 2025-01-08T10:45:43,895 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741905_1081 (size=65) 2025-01-08T10:45:43,896 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:45:43,896 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1681): Closing af16492a38d4fc130200d4b3697ff170, disabling compactions & flushes 2025-01-08T10:45:43,896 INFO [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170. 2025-01-08T10:45:43,896 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170. 2025-01-08T10:45:43,896 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170. after waiting 0 ms 2025-01-08T10:45:43,896 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170. 2025-01-08T10:45:43,896 INFO [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170. 2025-01-08T10:45:43,897 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1635): Region close journal for af16492a38d4fc130200d4b3697ff170: 2025-01-08T10:45:43,899 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741906_1082 (size=65) 2025-01-08T10:45:43,900 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741906_1082 (size=65) 2025-01-08T10:45:43,900 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741906_1082 (size=65) 2025-01-08T10:45:43,901 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:45:43,901 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1681): Closing 45518da4f8eaa6bc80e04f2fcdc6fc93, disabling compactions & flushes 2025-01-08T10:45:43,901 INFO [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93. 2025-01-08T10:45:43,901 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93. 2025-01-08T10:45:43,901 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93. after waiting 0 ms 2025-01-08T10:45:43,901 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93. 2025-01-08T10:45:43,901 INFO [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93. 2025-01-08T10:45:43,901 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1635): Region close journal for 45518da4f8eaa6bc80e04f2fcdc6fc93: 2025-01-08T10:45:43,903 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=30, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testExportWithResetTtl execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T10:45:43,903 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170.","families":{"info":[{"qualifier":"regioninfo","vlen":64,"tag":[],"timestamp":"1736333143903"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736333143903"}]},"ts":"1736333143903"} 2025-01-08T10:45:43,903 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93.","families":{"info":[{"qualifier":"regioninfo","vlen":64,"tag":[],"timestamp":"1736333143903"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736333143903"}]},"ts":"1736333143903"} 2025-01-08T10:45:43,907 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T10:45:43,908 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=30, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testExportWithResetTtl execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T10:45:43,908 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333143908"}]},"ts":"1736333143908"} 2025-01-08T10:45:43,910 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithResetTtl, state=ENABLING in hbase:meta 2025-01-08T10:45:43,914 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(202): Hosts are {67f42ee69b4e=0} racks are {/default-rack=0} 2025-01-08T10:45:43,916 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T10:45:43,916 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T10:45:43,916 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T10:45:43,916 INFO [PEWorker-2 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T10:45:43,916 INFO [PEWorker-2 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T10:45:43,916 INFO [PEWorker-2 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T10:45:43,916 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T10:45:43,916 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=31, ppid=30, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=af16492a38d4fc130200d4b3697ff170, ASSIGN}, {pid=32, ppid=30, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=45518da4f8eaa6bc80e04f2fcdc6fc93, ASSIGN}] 2025-01-08T10:45:43,918 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=32, ppid=30, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=45518da4f8eaa6bc80e04f2fcdc6fc93, ASSIGN 2025-01-08T10:45:43,918 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=31, ppid=30, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=af16492a38d4fc130200d4b3697ff170, ASSIGN 2025-01-08T10:45:43,919 INFO [PEWorker-5 {}] assignment.TransitRegionStateProcedure(264): Starting pid=32, ppid=30, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=45518da4f8eaa6bc80e04f2fcdc6fc93, ASSIGN; state=OFFLINE, location=67f42ee69b4e,45003,1736333108039; forceNewPlan=false, retain=false 2025-01-08T10:45:43,919 INFO [PEWorker-1 {}] assignment.TransitRegionStateProcedure(264): Starting pid=31, ppid=30, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=af16492a38d4fc130200d4b3697ff170, ASSIGN; state=OFFLINE, location=67f42ee69b4e,36283,1736333108166; forceNewPlan=false, retain=false 2025-01-08T10:45:43,973 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=30 2025-01-08T10:45:44,070 INFO [67f42ee69b4e:34551 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T10:45:44,071 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=32 updating hbase:meta row=45518da4f8eaa6bc80e04f2fcdc6fc93, regionState=OPENING, regionLocation=67f42ee69b4e,45003,1736333108039 2025-01-08T10:45:44,071 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=31 updating hbase:meta row=af16492a38d4fc130200d4b3697ff170, regionState=OPENING, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:44,073 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=33, ppid=32, state=RUNNABLE; OpenRegionProcedure 45518da4f8eaa6bc80e04f2fcdc6fc93, server=67f42ee69b4e,45003,1736333108039}] 2025-01-08T10:45:44,074 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=34, ppid=31, state=RUNNABLE; OpenRegionProcedure af16492a38d4fc130200d4b3697ff170, server=67f42ee69b4e,36283,1736333108166}] 2025-01-08T10:45:44,174 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=30 2025-01-08T10:45:44,225 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:45:44,227 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:44,229 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] handler.AssignRegionHandler(135): Open testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93. 2025-01-08T10:45:44,229 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(7285): Opening region: {ENCODED => 45518da4f8eaa6bc80e04f2fcdc6fc93, NAME => 'testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T10:45:44,230 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] handler.AssignRegionHandler(135): Open testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170. 2025-01-08T10:45:44,230 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(7285): Opening region: {ENCODED => af16492a38d4fc130200d4b3697ff170, NAME => 'testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T10:45:44,230 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93. service=AccessControlService 2025-01-08T10:45:44,230 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:45:44,230 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170. service=AccessControlService 2025-01-08T10:45:44,230 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportWithResetTtl 45518da4f8eaa6bc80e04f2fcdc6fc93 2025-01-08T10:45:44,231 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:45:44,231 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(894): Instantiated testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:45:44,231 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportWithResetTtl af16492a38d4fc130200d4b3697ff170 2025-01-08T10:45:44,231 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(7327): checking encryption for 45518da4f8eaa6bc80e04f2fcdc6fc93 2025-01-08T10:45:44,231 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(894): Instantiated testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:45:44,231 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(7330): checking classloading for 45518da4f8eaa6bc80e04f2fcdc6fc93 2025-01-08T10:45:44,231 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(7327): checking encryption for af16492a38d4fc130200d4b3697ff170 2025-01-08T10:45:44,231 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(7330): checking classloading for af16492a38d4fc130200d4b3697ff170 2025-01-08T10:45:44,232 INFO [StoreOpener-af16492a38d4fc130200d4b3697ff170-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region af16492a38d4fc130200d4b3697ff170 2025-01-08T10:45:44,232 INFO [StoreOpener-45518da4f8eaa6bc80e04f2fcdc6fc93-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 45518da4f8eaa6bc80e04f2fcdc6fc93 2025-01-08T10:45:44,234 INFO [StoreOpener-45518da4f8eaa6bc80e04f2fcdc6fc93-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 45518da4f8eaa6bc80e04f2fcdc6fc93 columnFamilyName cf 2025-01-08T10:45:44,235 INFO [StoreOpener-af16492a38d4fc130200d4b3697ff170-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region af16492a38d4fc130200d4b3697ff170 columnFamilyName cf 2025-01-08T10:45:44,235 DEBUG [StoreOpener-45518da4f8eaa6bc80e04f2fcdc6fc93-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:44,235 DEBUG [StoreOpener-af16492a38d4fc130200d4b3697ff170-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:44,236 INFO [StoreOpener-45518da4f8eaa6bc80e04f2fcdc6fc93-1 {}] regionserver.HStore(327): Store=45518da4f8eaa6bc80e04f2fcdc6fc93/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:45:44,236 INFO [StoreOpener-af16492a38d4fc130200d4b3697ff170-1 {}] regionserver.HStore(327): Store=af16492a38d4fc130200d4b3697ff170/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:45:44,237 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/af16492a38d4fc130200d4b3697ff170 2025-01-08T10:45:44,237 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/45518da4f8eaa6bc80e04f2fcdc6fc93 2025-01-08T10:45:44,237 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/af16492a38d4fc130200d4b3697ff170 2025-01-08T10:45:44,237 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/45518da4f8eaa6bc80e04f2fcdc6fc93 2025-01-08T10:45:44,240 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(1085): writing seq id for af16492a38d4fc130200d4b3697ff170 2025-01-08T10:45:44,240 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(1085): writing seq id for 45518da4f8eaa6bc80e04f2fcdc6fc93 2025-01-08T10:45:44,243 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/45518da4f8eaa6bc80e04f2fcdc6fc93/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:45:44,243 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/af16492a38d4fc130200d4b3697ff170/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:45:44,243 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(1102): Opened af16492a38d4fc130200d4b3697ff170; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=69358310, jitterRate=0.033519357442855835}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T10:45:44,244 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(1102): Opened 45518da4f8eaa6bc80e04f2fcdc6fc93; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=69489176, jitterRate=0.0354694128036499}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T10:45:44,244 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(1001): Region open journal for af16492a38d4fc130200d4b3697ff170: 2025-01-08T10:45:44,244 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(1001): Region open journal for 45518da4f8eaa6bc80e04f2fcdc6fc93: 2025-01-08T10:45:44,246 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93., pid=33, masterSystemTime=1736333144225 2025-01-08T10:45:44,246 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170., pid=34, masterSystemTime=1736333144226 2025-01-08T10:45:44,248 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93. 2025-01-08T10:45:44,248 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] handler.AssignRegionHandler(164): Opened testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93. 2025-01-08T10:45:44,249 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=32 updating hbase:meta row=45518da4f8eaa6bc80e04f2fcdc6fc93, regionState=OPEN, openSeqNum=2, regionLocation=67f42ee69b4e,45003,1736333108039 2025-01-08T10:45:44,249 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170. 2025-01-08T10:45:44,249 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] handler.AssignRegionHandler(164): Opened testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170. 2025-01-08T10:45:44,250 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=31 updating hbase:meta row=af16492a38d4fc130200d4b3697ff170, regionState=OPEN, openSeqNum=2, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:44,254 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=33, resume processing ppid=32 2025-01-08T10:45:44,254 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=33, ppid=32, state=SUCCESS; OpenRegionProcedure 45518da4f8eaa6bc80e04f2fcdc6fc93, server=67f42ee69b4e,45003,1736333108039 in 178 msec 2025-01-08T10:45:44,255 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=34, resume processing ppid=31 2025-01-08T10:45:44,255 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=34, ppid=31, state=SUCCESS; OpenRegionProcedure af16492a38d4fc130200d4b3697ff170, server=67f42ee69b4e,36283,1736333108166 in 178 msec 2025-01-08T10:45:44,256 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=32, ppid=30, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=45518da4f8eaa6bc80e04f2fcdc6fc93, ASSIGN in 338 msec 2025-01-08T10:45:44,257 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=31, resume processing ppid=30 2025-01-08T10:45:44,257 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=31, ppid=30, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=af16492a38d4fc130200d4b3697ff170, ASSIGN in 339 msec 2025-01-08T10:45:44,258 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=30, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testExportWithResetTtl execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T10:45:44,258 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333144258"}]},"ts":"1736333144258"} 2025-01-08T10:45:44,260 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithResetTtl, state=ENABLED in hbase:meta 2025-01-08T10:45:44,262 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=30, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportWithResetTtl execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T10:45:44,263 DEBUG [PEWorker-1 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testExportWithResetTtl jenkins: RWXCA 2025-01-08T10:45:44,265 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36283 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportWithResetTtl], kv [jenkins: RWXCA] 2025-01-08T10:45:44,267 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:45:44,267 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:45:44,267 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:45:44,267 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:45:44,269 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T10:45:44,269 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T10:45:44,269 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T10:45:44,269 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T10:45:44,270 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=30, state=SUCCESS; CreateTableProcedure table=testtb-testExportWithResetTtl in 401 msec 2025-01-08T10:45:44,476 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=30 2025-01-08T10:45:44,477 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testExportWithResetTtl, procId: 30 completed 2025-01-08T10:45:44,480 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportWithResetTtl 2025-01-08T10:45:44,480 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170. 2025-01-08T10:45:44,480 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T10:45:44,496 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.2 snapshot request for:{ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } 2025-01-08T10:45:44,496 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736333144496 (current time:1736333144496). 2025-01-08T10:45:44,496 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T10:45:44,496 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot emptySnaptb0-testExportWithResetTtl VERSION not specified, setting to 2 2025-01-08T10:45:44,496 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T10:45:44,497 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x792c0066 to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@70df6834 2025-01-08T10:45:44,501 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5a90a305, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:45:44,502 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:45:44,503 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:50976, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:45:44,504 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x792c0066 to 127.0.0.1:52367 2025-01-08T10:45:44,505 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:45:44,506 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x1487c378 to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@1f928714 2025-01-08T10:45:44,511 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@40458768, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:45:44,513 DEBUG [hconnection-0x439b5ff5-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:45:44,514 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:50990, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:45:44,517 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x1487c378 to 127.0.0.1:52367 2025-01-08T10:45:44,517 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:45:44,517 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportWithResetTtl], kv [jenkins: RWXCA] 2025-01-08T10:45:44,518 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T10:45:44,519 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=35, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } 2025-01-08T10:45:44,519 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 }, snapshot procedure id = 35 2025-01-08T10:45:44,520 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T10:45:44,520 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=35 2025-01-08T10:45:44,521 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T10:45:44,524 DEBUG [BootstrapNodeManager {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2025-01-08T10:45:44,525 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T10:45:44,525 INFO [RS-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:59224, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.2 (auth:SIMPLE), service=RegionServerStatusService 2025-01-08T10:45:44,533 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741907_1083 (size=161) 2025-01-08T10:45:44,533 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741907_1083 (size=161) 2025-01-08T10:45:44,534 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741907_1083 (size=161) 2025-01-08T10:45:44,537 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T10:45:44,537 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=36, ppid=35, state=RUNNABLE; SnapshotRegionProcedure af16492a38d4fc130200d4b3697ff170}, {pid=37, ppid=35, state=RUNNABLE; SnapshotRegionProcedure 45518da4f8eaa6bc80e04f2fcdc6fc93}] 2025-01-08T10:45:44,538 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=37, ppid=35, state=RUNNABLE; SnapshotRegionProcedure 45518da4f8eaa6bc80e04f2fcdc6fc93 2025-01-08T10:45:44,538 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=36, ppid=35, state=RUNNABLE; SnapshotRegionProcedure af16492a38d4fc130200d4b3697ff170 2025-01-08T10:45:44,621 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=35 2025-01-08T10:45:44,689 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:45:44,689 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:44,690 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=45003 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=37 2025-01-08T10:45:44,690 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=36283 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=36 2025-01-08T10:45:44,692 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=37}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93. 2025-01-08T10:45:44,693 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=36}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170. 2025-01-08T10:45:44,693 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=37}] regionserver.HRegion(2538): Flush status journal for 45518da4f8eaa6bc80e04f2fcdc6fc93: 2025-01-08T10:45:44,693 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=36}] regionserver.HRegion(2538): Flush status journal for af16492a38d4fc130200d4b3697ff170: 2025-01-08T10:45:44,693 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=37}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93. for emptySnaptb0-testExportWithResetTtl completed. 2025-01-08T10:45:44,693 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=36}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170. for emptySnaptb0-testExportWithResetTtl completed. 2025-01-08T10:45:44,694 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=37}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93.' region-info for snapshot=emptySnaptb0-testExportWithResetTtl 2025-01-08T10:45:44,694 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=36}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170.' region-info for snapshot=emptySnaptb0-testExportWithResetTtl 2025-01-08T10:45:44,694 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=37}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:45:44,694 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=36}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:45:44,694 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=37}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T10:45:44,694 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=36}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T10:45:44,713 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741909_1085 (size=68) 2025-01-08T10:45:44,714 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741909_1085 (size=68) 2025-01-08T10:45:44,715 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741909_1085 (size=68) 2025-01-08T10:45:44,716 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741908_1084 (size=68) 2025-01-08T10:45:44,716 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741908_1084 (size=68) 2025-01-08T10:45:44,716 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=37}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93. 2025-01-08T10:45:44,716 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=37}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=37 2025-01-08T10:45:44,716 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=37 2025-01-08T10:45:44,717 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportWithResetTtl on region 45518da4f8eaa6bc80e04f2fcdc6fc93 2025-01-08T10:45:44,717 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=37, ppid=35, state=RUNNABLE; SnapshotRegionProcedure 45518da4f8eaa6bc80e04f2fcdc6fc93 2025-01-08T10:45:44,719 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741908_1084 (size=68) 2025-01-08T10:45:44,719 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=36}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170. 2025-01-08T10:45:44,719 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=36}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=36 2025-01-08T10:45:44,720 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=36 2025-01-08T10:45:44,720 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportWithResetTtl on region af16492a38d4fc130200d4b3697ff170 2025-01-08T10:45:44,720 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=36, ppid=35, state=RUNNABLE; SnapshotRegionProcedure af16492a38d4fc130200d4b3697ff170 2025-01-08T10:45:44,721 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=37, ppid=35, state=SUCCESS; SnapshotRegionProcedure 45518da4f8eaa6bc80e04f2fcdc6fc93 in 181 msec 2025-01-08T10:45:44,725 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=36, resume processing ppid=35 2025-01-08T10:45:44,725 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=36, ppid=35, state=SUCCESS; SnapshotRegionProcedure af16492a38d4fc130200d4b3697ff170 in 185 msec 2025-01-08T10:45:44,725 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T10:45:44,726 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T10:45:44,728 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T10:45:44,728 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T10:45:44,728 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:44,729 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(328): No files under family: cf 2025-01-08T10:45:44,753 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741910_1086 (size=60) 2025-01-08T10:45:44,754 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741910_1086 (size=60) 2025-01-08T10:45:44,754 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741910_1086 (size=60) 2025-01-08T10:45:44,757 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T10:45:44,757 DEBUG [PEWorker-4 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for emptySnaptb0-testExportWithResetTtl 2025-01-08T10:45:44,758 DEBUG [PEWorker-4 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/emptySnaptb0-testExportWithResetTtl 2025-01-08T10:45:44,781 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741911_1087 (size=641) 2025-01-08T10:45:44,781 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741911_1087 (size=641) 2025-01-08T10:45:44,782 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741911_1087 (size=641) 2025-01-08T10:45:44,786 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T10:45:44,793 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T10:45:44,793 DEBUG [PEWorker-4 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/emptySnaptb0-testExportWithResetTtl to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/emptySnaptb0-testExportWithResetTtl 2025-01-08T10:45:44,795 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T10:45:44,795 DEBUG [PEWorker-4 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 }, snapshot procedure id = 35 2025-01-08T10:45:44,797 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=35, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } in 277 msec 2025-01-08T10:45:44,823 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=35 2025-01-08T10:45:44,823 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportWithResetTtl, procId: 35 completed 2025-01-08T10:45:44,827 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36283 {}] regionserver.HRegion(8254): writing data to region testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T10:45:44,830 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=45003 {}] regionserver.HRegion(8254): writing data to region testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T10:45:44,837 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportWithResetTtl 2025-01-08T10:45:44,837 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170. 2025-01-08T10:45:44,837 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T10:45:44,852 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.2 snapshot request for:{ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } 2025-01-08T10:45:44,852 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736333144852 (current time:1736333144852). 2025-01-08T10:45:44,853 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T10:45:44,853 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testExportWithResetTtl VERSION not specified, setting to 2 2025-01-08T10:45:44,853 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T10:45:44,854 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x0e9eed05 to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@17637fd4 2025-01-08T10:45:44,858 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@100ea7e5, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:45:44,861 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:45:44,863 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:50992, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:45:44,864 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x0e9eed05 to 127.0.0.1:52367 2025-01-08T10:45:44,864 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:45:44,866 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x000fb2b4 to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@3b0515bc 2025-01-08T10:45:44,870 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@400c2aab, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:45:44,873 DEBUG [hconnection-0x53b77fe3-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:45:44,874 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:51008, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:45:44,878 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x000fb2b4 to 127.0.0.1:52367 2025-01-08T10:45:44,878 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:45:44,878 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportWithResetTtl], kv [jenkins: RWXCA] 2025-01-08T10:45:44,879 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T10:45:44,880 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=38, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } 2025-01-08T10:45:44,880 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 }, snapshot procedure id = 38 2025-01-08T10:45:44,881 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T10:45:44,882 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=38 2025-01-08T10:45:44,882 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T10:45:44,885 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T10:45:44,896 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741912_1088 (size=156) 2025-01-08T10:45:44,897 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741912_1088 (size=156) 2025-01-08T10:45:44,897 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741912_1088 (size=156) 2025-01-08T10:45:44,898 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T10:45:44,899 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=39, ppid=38, state=RUNNABLE; SnapshotRegionProcedure af16492a38d4fc130200d4b3697ff170}, {pid=40, ppid=38, state=RUNNABLE; SnapshotRegionProcedure 45518da4f8eaa6bc80e04f2fcdc6fc93}] 2025-01-08T10:45:44,900 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=39, ppid=38, state=RUNNABLE; SnapshotRegionProcedure af16492a38d4fc130200d4b3697ff170 2025-01-08T10:45:44,900 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=40, ppid=38, state=RUNNABLE; SnapshotRegionProcedure 45518da4f8eaa6bc80e04f2fcdc6fc93 2025-01-08T10:45:44,983 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=38 2025-01-08T10:45:45,051 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:45:45,051 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:45,052 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=36283 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=39 2025-01-08T10:45:45,052 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=45003 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=40 2025-01-08T10:45:45,052 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93. 2025-01-08T10:45:45,052 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170. 2025-01-08T10:45:45,052 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.HRegion(2837): Flushing 45518da4f8eaa6bc80e04f2fcdc6fc93 1/1 column families, dataSize=3.13 KB heapSize=7 KB 2025-01-08T10:45:45,052 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.HRegion(2837): Flushing af16492a38d4fc130200d4b3697ff170 1/1 column families, dataSize=132 B heapSize=544 B 2025-01-08T10:45:45,075 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108089b7e8178ce411e861aa3c3847e56cf_45518da4f8eaa6bc80e04f2fcdc6fc93 is 71, key is 155f5b66d68e3f5407d5e1f6aa055812/cf:q/1736333144830/Put/seqid=0 2025-01-08T10:45:45,078 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108870de8e9e6324e44bef2c74945990200_af16492a38d4fc130200d4b3697ff170 is 71, key is 0ab88f02ee864370b9c60d3d1c3f6938/cf:q/1736333144827/Put/seqid=0 2025-01-08T10:45:45,085 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741913_1089 (size=8241) 2025-01-08T10:45:45,086 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741913_1089 (size=8241) 2025-01-08T10:45:45,086 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741913_1089 (size=8241) 2025-01-08T10:45:45,086 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:45,088 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741914_1090 (size=5032) 2025-01-08T10:45:45,091 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741914_1090 (size=5032) 2025-01-08T10:45:45,091 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741914_1090 (size=5032) 2025-01-08T10:45:45,091 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:45,096 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108089b7e8178ce411e861aa3c3847e56cf_45518da4f8eaa6bc80e04f2fcdc6fc93 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/c4ca4238a0b923820dcc509a6f75849b20250108089b7e8178ce411e861aa3c3847e56cf_45518da4f8eaa6bc80e04f2fcdc6fc93 2025-01-08T10:45:45,098 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/45518da4f8eaa6bc80e04f2fcdc6fc93/.tmp/cf/d4088f10677346b5b0f9e22c32824a43, store: [table=testtb-testExportWithResetTtl family=cf region=45518da4f8eaa6bc80e04f2fcdc6fc93] 2025-01-08T10:45:45,099 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/45518da4f8eaa6bc80e04f2fcdc6fc93/.tmp/cf/d4088f10677346b5b0f9e22c32824a43 is 206, key is 1f5a40780d1b79c4c2f53cb7968bf9cab/cf:q/1736333144830/Put/seqid=0 2025-01-08T10:45:45,099 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108870de8e9e6324e44bef2c74945990200_af16492a38d4fc130200d4b3697ff170 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/d41d8cd98f00b204e9800998ecf8427e20250108870de8e9e6324e44bef2c74945990200_af16492a38d4fc130200d4b3697ff170 2025-01-08T10:45:45,100 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/af16492a38d4fc130200d4b3697ff170/.tmp/cf/0ea6716cd7184a7f9071648966df2e6c, store: [table=testtb-testExportWithResetTtl family=cf region=af16492a38d4fc130200d4b3697ff170] 2025-01-08T10:45:45,102 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/af16492a38d4fc130200d4b3697ff170/.tmp/cf/0ea6716cd7184a7f9071648966df2e6c is 206, key is 027b2dc24b0e6128a5a12102e306a2d33/cf:q/1736333144827/Put/seqid=0 2025-01-08T10:45:45,113 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741915_1091 (size=15055) 2025-01-08T10:45:45,118 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741915_1091 (size=15055) 2025-01-08T10:45:45,118 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741915_1091 (size=15055) 2025-01-08T10:45:45,119 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=3.1 K, hasBloomFilter=true, into tmp file hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/45518da4f8eaa6bc80e04f2fcdc6fc93/.tmp/cf/d4088f10677346b5b0f9e22c32824a43 2025-01-08T10:45:45,120 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741916_1092 (size=5700) 2025-01-08T10:45:45,121 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741916_1092 (size=5700) 2025-01-08T10:45:45,126 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741916_1092 (size=5700) 2025-01-08T10:45:45,127 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=132, hasBloomFilter=true, into tmp file hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/af16492a38d4fc130200d4b3697ff170/.tmp/cf/0ea6716cd7184a7f9071648966df2e6c 2025-01-08T10:45:45,128 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/45518da4f8eaa6bc80e04f2fcdc6fc93/.tmp/cf/d4088f10677346b5b0f9e22c32824a43 as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/45518da4f8eaa6bc80e04f2fcdc6fc93/cf/d4088f10677346b5b0f9e22c32824a43 2025-01-08T10:45:45,136 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/45518da4f8eaa6bc80e04f2fcdc6fc93/cf/d4088f10677346b5b0f9e22c32824a43, entries=48, sequenceid=6, filesize=14.7 K 2025-01-08T10:45:45,136 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/af16492a38d4fc130200d4b3697ff170/.tmp/cf/0ea6716cd7184a7f9071648966df2e6c as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/af16492a38d4fc130200d4b3697ff170/cf/0ea6716cd7184a7f9071648966df2e6c 2025-01-08T10:45:45,137 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.HRegion(3040): Finished flush of dataSize ~3.13 KB/3204, heapSize ~6.98 KB/7152, currentSize=0 B/0 for 45518da4f8eaa6bc80e04f2fcdc6fc93 in 85ms, sequenceid=6, compaction requested=false 2025-01-08T10:45:45,137 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportWithResetTtl' 2025-01-08T10:45:45,138 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.HRegion(2538): Flush status journal for 45518da4f8eaa6bc80e04f2fcdc6fc93: 2025-01-08T10:45:45,138 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93. for snaptb0-testExportWithResetTtl completed. 2025-01-08T10:45:45,138 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93.' region-info for snapshot=snaptb0-testExportWithResetTtl 2025-01-08T10:45:45,138 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:45:45,138 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/45518da4f8eaa6bc80e04f2fcdc6fc93/cf/d4088f10677346b5b0f9e22c32824a43] hfiles 2025-01-08T10:45:45,138 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/45518da4f8eaa6bc80e04f2fcdc6fc93/cf/d4088f10677346b5b0f9e22c32824a43 for snapshot=snaptb0-testExportWithResetTtl 2025-01-08T10:45:45,152 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/af16492a38d4fc130200d4b3697ff170/cf/0ea6716cd7184a7f9071648966df2e6c, entries=2, sequenceid=6, filesize=5.6 K 2025-01-08T10:45:45,162 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.HRegion(3040): Finished flush of dataSize ~132 B/132, heapSize ~528 B/528, currentSize=0 B/0 for af16492a38d4fc130200d4b3697ff170 in 110ms, sequenceid=6, compaction requested=false 2025-01-08T10:45:45,162 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.HRegion(2538): Flush status journal for af16492a38d4fc130200d4b3697ff170: 2025-01-08T10:45:45,162 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170. for snaptb0-testExportWithResetTtl completed. 2025-01-08T10:45:45,163 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170.' region-info for snapshot=snaptb0-testExportWithResetTtl 2025-01-08T10:45:45,163 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:45:45,163 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/af16492a38d4fc130200d4b3697ff170/cf/0ea6716cd7184a7f9071648966df2e6c] hfiles 2025-01-08T10:45:45,163 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/af16492a38d4fc130200d4b3697ff170/cf/0ea6716cd7184a7f9071648966df2e6c for snapshot=snaptb0-testExportWithResetTtl 2025-01-08T10:45:45,173 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741917_1093 (size=107) 2025-01-08T10:45:45,173 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741917_1093 (size=107) 2025-01-08T10:45:45,173 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741917_1093 (size=107) 2025-01-08T10:45:45,174 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93. 2025-01-08T10:45:45,174 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=40 2025-01-08T10:45:45,174 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=40 2025-01-08T10:45:45,174 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportWithResetTtl on region 45518da4f8eaa6bc80e04f2fcdc6fc93 2025-01-08T10:45:45,175 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=40, ppid=38, state=RUNNABLE; SnapshotRegionProcedure 45518da4f8eaa6bc80e04f2fcdc6fc93 2025-01-08T10:45:45,177 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=40, ppid=38, state=SUCCESS; SnapshotRegionProcedure 45518da4f8eaa6bc80e04f2fcdc6fc93 in 277 msec 2025-01-08T10:45:45,184 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=38 2025-01-08T10:45:45,188 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741918_1094 (size=107) 2025-01-08T10:45:45,188 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741918_1094 (size=107) 2025-01-08T10:45:45,189 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741918_1094 (size=107) 2025-01-08T10:45:45,189 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170. 2025-01-08T10:45:45,190 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=39 2025-01-08T10:45:45,190 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=39 2025-01-08T10:45:45,190 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportWithResetTtl on region af16492a38d4fc130200d4b3697ff170 2025-01-08T10:45:45,190 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=39, ppid=38, state=RUNNABLE; SnapshotRegionProcedure af16492a38d4fc130200d4b3697ff170 2025-01-08T10:45:45,193 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=39, resume processing ppid=38 2025-01-08T10:45:45,193 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T10:45:45,193 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=39, ppid=38, state=SUCCESS; SnapshotRegionProcedure af16492a38d4fc130200d4b3697ff170 in 292 msec 2025-01-08T10:45:45,194 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T10:45:45,196 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T10:45:45,196 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T10:45:45,196 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:45,197 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/c4ca4238a0b923820dcc509a6f75849b20250108089b7e8178ce411e861aa3c3847e56cf_45518da4f8eaa6bc80e04f2fcdc6fc93, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/d41d8cd98f00b204e9800998ecf8427e20250108870de8e9e6324e44bef2c74945990200_af16492a38d4fc130200d4b3697ff170] hfiles 2025-01-08T10:45:45,197 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/c4ca4238a0b923820dcc509a6f75849b20250108089b7e8178ce411e861aa3c3847e56cf_45518da4f8eaa6bc80e04f2fcdc6fc93 2025-01-08T10:45:45,197 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/d41d8cd98f00b204e9800998ecf8427e20250108870de8e9e6324e44bef2c74945990200_af16492a38d4fc130200d4b3697ff170 2025-01-08T10:45:45,204 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741919_1095 (size=291) 2025-01-08T10:45:45,205 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741919_1095 (size=291) 2025-01-08T10:45:45,205 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741919_1095 (size=291) 2025-01-08T10:45:45,207 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T10:45:45,207 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testExportWithResetTtl 2025-01-08T10:45:45,208 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/snaptb0-testExportWithResetTtl 2025-01-08T10:45:45,218 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741920_1096 (size=951) 2025-01-08T10:45:45,218 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741920_1096 (size=951) 2025-01-08T10:45:45,219 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741920_1096 (size=951) 2025-01-08T10:45:45,225 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T10:45:45,232 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T10:45:45,233 DEBUG [PEWorker-3 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/snaptb0-testExportWithResetTtl to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportWithResetTtl 2025-01-08T10:45:45,234 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T10:45:45,234 DEBUG [PEWorker-3 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 }, snapshot procedure id = 38 2025-01-08T10:45:45,236 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=38, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } in 356 msec 2025-01-08T10:45:45,486 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=38 2025-01-08T10:45:45,486 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportWithResetTtl, procId: 38 completed 2025-01-08T10:45:45,488 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.2 create 'testExportWithResetTtl', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T10:45:45,490 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=41, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testExportWithResetTtl 2025-01-08T10:45:45,491 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=41, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testExportWithResetTtl execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T10:45:45,491 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "testExportWithResetTtl" procId is: 41 2025-01-08T10:45:45,492 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=41, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testExportWithResetTtl execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T10:45:45,492 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=41 2025-01-08T10:45:45,506 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741921_1097 (size=433) 2025-01-08T10:45:45,507 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741921_1097 (size=433) 2025-01-08T10:45:45,507 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741921_1097 (size=433) 2025-01-08T10:45:45,510 INFO [RegionOpenAndInit-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => d553a3cb3f214b19b4267c8eb74194a5, NAME => 'testExportWithResetTtl,1,1736333145488.d553a3cb3f214b19b4267c8eb74194a5.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testExportWithResetTtl', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:45:45,510 INFO [RegionOpenAndInit-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 053a5459e246ced75c7eec0a1d77ff06, NAME => 'testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testExportWithResetTtl', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:45:45,533 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741922_1098 (size=58) 2025-01-08T10:45:45,533 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741922_1098 (size=58) 2025-01-08T10:45:45,533 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741922_1098 (size=58) 2025-01-08T10:45:45,536 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741923_1099 (size=58) 2025-01-08T10:45:45,536 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741923_1099 (size=58) 2025-01-08T10:45:45,536 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741923_1099 (size=58) 2025-01-08T10:45:45,538 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(894): Instantiated testExportWithResetTtl,1,1736333145488.d553a3cb3f214b19b4267c8eb74194a5.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:45:45,538 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1681): Closing d553a3cb3f214b19b4267c8eb74194a5, disabling compactions & flushes 2025-01-08T10:45:45,539 INFO [RegionOpenAndInit-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1703): Closing region testExportWithResetTtl,1,1736333145488.d553a3cb3f214b19b4267c8eb74194a5. 2025-01-08T10:45:45,539 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testExportWithResetTtl,1,1736333145488.d553a3cb3f214b19b4267c8eb74194a5. 2025-01-08T10:45:45,539 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testExportWithResetTtl,1,1736333145488.d553a3cb3f214b19b4267c8eb74194a5. after waiting 0 ms 2025-01-08T10:45:45,539 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testExportWithResetTtl,1,1736333145488.d553a3cb3f214b19b4267c8eb74194a5. 2025-01-08T10:45:45,539 INFO [RegionOpenAndInit-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1922): Closed testExportWithResetTtl,1,1736333145488.d553a3cb3f214b19b4267c8eb74194a5. 2025-01-08T10:45:45,539 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1635): Region close journal for d553a3cb3f214b19b4267c8eb74194a5: 2025-01-08T10:45:45,594 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=41 2025-01-08T10:45:45,795 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=41 2025-01-08T10:45:45,934 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(894): Instantiated testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:45:45,934 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1681): Closing 053a5459e246ced75c7eec0a1d77ff06, disabling compactions & flushes 2025-01-08T10:45:45,934 INFO [RegionOpenAndInit-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1703): Closing region testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06. 2025-01-08T10:45:45,934 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06. 2025-01-08T10:45:45,934 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06. after waiting 0 ms 2025-01-08T10:45:45,934 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06. 2025-01-08T10:45:45,934 INFO [RegionOpenAndInit-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1922): Closed testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06. 2025-01-08T10:45:45,935 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1635): Region close journal for 053a5459e246ced75c7eec0a1d77ff06: 2025-01-08T10:45:45,936 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=41, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testExportWithResetTtl execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T10:45:45,936 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testExportWithResetTtl,1,1736333145488.d553a3cb3f214b19b4267c8eb74194a5.","families":{"info":[{"qualifier":"regioninfo","vlen":57,"tag":[],"timestamp":"1736333145936"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736333145936"}]},"ts":"1736333145936"} 2025-01-08T10:45:45,937 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06.","families":{"info":[{"qualifier":"regioninfo","vlen":57,"tag":[],"timestamp":"1736333145936"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736333145936"}]},"ts":"1736333145936"} 2025-01-08T10:45:45,939 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T10:45:45,940 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=41, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testExportWithResetTtl execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T10:45:45,940 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333145940"}]},"ts":"1736333145940"} 2025-01-08T10:45:45,942 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=testExportWithResetTtl, state=ENABLING in hbase:meta 2025-01-08T10:45:45,946 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(202): Hosts are {67f42ee69b4e=0} racks are {/default-rack=0} 2025-01-08T10:45:45,947 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T10:45:45,947 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T10:45:45,948 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T10:45:45,948 INFO [PEWorker-5 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T10:45:45,948 INFO [PEWorker-5 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T10:45:45,948 INFO [PEWorker-5 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T10:45:45,948 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T10:45:45,948 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=42, ppid=41, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testExportWithResetTtl, region=053a5459e246ced75c7eec0a1d77ff06, ASSIGN}, {pid=43, ppid=41, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testExportWithResetTtl, region=d553a3cb3f214b19b4267c8eb74194a5, ASSIGN}] 2025-01-08T10:45:45,949 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=43, ppid=41, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testExportWithResetTtl, region=d553a3cb3f214b19b4267c8eb74194a5, ASSIGN 2025-01-08T10:45:45,949 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=42, ppid=41, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testExportWithResetTtl, region=053a5459e246ced75c7eec0a1d77ff06, ASSIGN 2025-01-08T10:45:45,950 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=43, ppid=41, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testExportWithResetTtl, region=d553a3cb3f214b19b4267c8eb74194a5, ASSIGN; state=OFFLINE, location=67f42ee69b4e,36283,1736333108166; forceNewPlan=false, retain=false 2025-01-08T10:45:45,950 INFO [PEWorker-1 {}] assignment.TransitRegionStateProcedure(264): Starting pid=42, ppid=41, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testExportWithResetTtl, region=053a5459e246ced75c7eec0a1d77ff06, ASSIGN; state=OFFLINE, location=67f42ee69b4e,45003,1736333108039; forceNewPlan=false, retain=false 2025-01-08T10:45:46,097 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=41 2025-01-08T10:45:46,100 INFO [67f42ee69b4e:34551 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T10:45:46,100 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=43 updating hbase:meta row=d553a3cb3f214b19b4267c8eb74194a5, regionState=OPENING, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:46,100 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=42 updating hbase:meta row=053a5459e246ced75c7eec0a1d77ff06, regionState=OPENING, regionLocation=67f42ee69b4e,45003,1736333108039 2025-01-08T10:45:46,102 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=44, ppid=43, state=RUNNABLE; OpenRegionProcedure d553a3cb3f214b19b4267c8eb74194a5, server=67f42ee69b4e,36283,1736333108166}] 2025-01-08T10:45:46,103 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=45, ppid=42, state=RUNNABLE; OpenRegionProcedure 053a5459e246ced75c7eec0a1d77ff06, server=67f42ee69b4e,45003,1736333108039}] 2025-01-08T10:45:46,255 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:46,256 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:45:46,259 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] handler.AssignRegionHandler(135): Open testExportWithResetTtl,1,1736333145488.d553a3cb3f214b19b4267c8eb74194a5. 2025-01-08T10:45:46,259 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(7285): Opening region: {ENCODED => d553a3cb3f214b19b4267c8eb74194a5, NAME => 'testExportWithResetTtl,1,1736333145488.d553a3cb3f214b19b4267c8eb74194a5.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T10:45:46,260 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] handler.AssignRegionHandler(135): Open testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06. 2025-01-08T10:45:46,260 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(7285): Opening region: {ENCODED => 053a5459e246ced75c7eec0a1d77ff06, NAME => 'testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T10:45:46,260 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(7999): Registered coprocessor service: region=testExportWithResetTtl,1,1736333145488.d553a3cb3f214b19b4267c8eb74194a5. service=AccessControlService 2025-01-08T10:45:46,260 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(7999): Registered coprocessor service: region=testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06. service=AccessControlService 2025-01-08T10:45:46,260 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:45:46,260 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:45:46,260 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testExportWithResetTtl 053a5459e246ced75c7eec0a1d77ff06 2025-01-08T10:45:46,260 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testExportWithResetTtl d553a3cb3f214b19b4267c8eb74194a5 2025-01-08T10:45:46,260 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(894): Instantiated testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:45:46,260 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(894): Instantiated testExportWithResetTtl,1,1736333145488.d553a3cb3f214b19b4267c8eb74194a5.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:45:46,261 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(7327): checking encryption for 053a5459e246ced75c7eec0a1d77ff06 2025-01-08T10:45:46,261 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(7330): checking classloading for 053a5459e246ced75c7eec0a1d77ff06 2025-01-08T10:45:46,261 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(7327): checking encryption for d553a3cb3f214b19b4267c8eb74194a5 2025-01-08T10:45:46,261 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(7330): checking classloading for d553a3cb3f214b19b4267c8eb74194a5 2025-01-08T10:45:46,262 INFO [StoreOpener-053a5459e246ced75c7eec0a1d77ff06-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 053a5459e246ced75c7eec0a1d77ff06 2025-01-08T10:45:46,262 INFO [StoreOpener-d553a3cb3f214b19b4267c8eb74194a5-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region d553a3cb3f214b19b4267c8eb74194a5 2025-01-08T10:45:46,264 INFO [StoreOpener-053a5459e246ced75c7eec0a1d77ff06-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 053a5459e246ced75c7eec0a1d77ff06 columnFamilyName cf 2025-01-08T10:45:46,265 DEBUG [StoreOpener-053a5459e246ced75c7eec0a1d77ff06-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:46,266 INFO [StoreOpener-053a5459e246ced75c7eec0a1d77ff06-1 {}] regionserver.HStore(327): Store=053a5459e246ced75c7eec0a1d77ff06/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:45:46,267 INFO [StoreOpener-d553a3cb3f214b19b4267c8eb74194a5-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region d553a3cb3f214b19b4267c8eb74194a5 columnFamilyName cf 2025-01-08T10:45:46,267 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/053a5459e246ced75c7eec0a1d77ff06 2025-01-08T10:45:46,267 DEBUG [StoreOpener-d553a3cb3f214b19b4267c8eb74194a5-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:46,267 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/053a5459e246ced75c7eec0a1d77ff06 2025-01-08T10:45:46,268 INFO [StoreOpener-d553a3cb3f214b19b4267c8eb74194a5-1 {}] regionserver.HStore(327): Store=d553a3cb3f214b19b4267c8eb74194a5/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:45:46,269 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/d553a3cb3f214b19b4267c8eb74194a5 2025-01-08T10:45:46,270 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/d553a3cb3f214b19b4267c8eb74194a5 2025-01-08T10:45:46,270 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(1085): writing seq id for 053a5459e246ced75c7eec0a1d77ff06 2025-01-08T10:45:46,273 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(1085): writing seq id for d553a3cb3f214b19b4267c8eb74194a5 2025-01-08T10:45:46,274 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/053a5459e246ced75c7eec0a1d77ff06/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:45:46,274 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(1102): Opened 053a5459e246ced75c7eec0a1d77ff06; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=72182146, jitterRate=0.07559779286384583}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T10:45:46,276 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(1001): Region open journal for 053a5459e246ced75c7eec0a1d77ff06: 2025-01-08T10:45:46,277 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/d553a3cb3f214b19b4267c8eb74194a5/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:45:46,277 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegionServer(2601): Post open deploy tasks for testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06., pid=45, masterSystemTime=1736333146256 2025-01-08T10:45:46,277 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(1102): Opened d553a3cb3f214b19b4267c8eb74194a5; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=69102913, jitterRate=0.029713645577430725}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T10:45:46,277 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(1001): Region open journal for d553a3cb3f214b19b4267c8eb74194a5: 2025-01-08T10:45:46,279 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegionServer(2628): Finished post open deploy task for testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06. 2025-01-08T10:45:46,279 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] handler.AssignRegionHandler(164): Opened testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06. 2025-01-08T10:45:46,280 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=42 updating hbase:meta row=053a5459e246ced75c7eec0a1d77ff06, regionState=OPEN, openSeqNum=2, regionLocation=67f42ee69b4e,45003,1736333108039 2025-01-08T10:45:46,282 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegionServer(2601): Post open deploy tasks for testExportWithResetTtl,1,1736333145488.d553a3cb3f214b19b4267c8eb74194a5., pid=44, masterSystemTime=1736333146255 2025-01-08T10:45:46,284 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegionServer(2628): Finished post open deploy task for testExportWithResetTtl,1,1736333145488.d553a3cb3f214b19b4267c8eb74194a5. 2025-01-08T10:45:46,284 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] handler.AssignRegionHandler(164): Opened testExportWithResetTtl,1,1736333145488.d553a3cb3f214b19b4267c8eb74194a5. 2025-01-08T10:45:46,285 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=43 updating hbase:meta row=d553a3cb3f214b19b4267c8eb74194a5, regionState=OPEN, openSeqNum=2, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:46,286 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=45, resume processing ppid=42 2025-01-08T10:45:46,286 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=45, ppid=42, state=SUCCESS; OpenRegionProcedure 053a5459e246ced75c7eec0a1d77ff06, server=67f42ee69b4e,45003,1736333108039 in 180 msec 2025-01-08T10:45:46,288 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=42, ppid=41, state=SUCCESS; TransitRegionStateProcedure table=testExportWithResetTtl, region=053a5459e246ced75c7eec0a1d77ff06, ASSIGN in 338 msec 2025-01-08T10:45:46,289 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=44, resume processing ppid=43 2025-01-08T10:45:46,289 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=44, ppid=43, state=SUCCESS; OpenRegionProcedure d553a3cb3f214b19b4267c8eb74194a5, server=67f42ee69b4e,36283,1736333108166 in 185 msec 2025-01-08T10:45:46,291 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=43, resume processing ppid=41 2025-01-08T10:45:46,291 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=43, ppid=41, state=SUCCESS; TransitRegionStateProcedure table=testExportWithResetTtl, region=d553a3cb3f214b19b4267c8eb74194a5, ASSIGN in 341 msec 2025-01-08T10:45:46,292 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=41, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testExportWithResetTtl execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T10:45:46,292 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333146292"}]},"ts":"1736333146292"} 2025-01-08T10:45:46,293 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testExportWithResetTtl, state=ENABLED in hbase:meta 2025-01-08T10:45:46,296 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=41, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testExportWithResetTtl execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T10:45:46,297 DEBUG [PEWorker-2 {}] access.PermissionStorage(175): Writing permission with rowKey testExportWithResetTtl jenkins: RWXCA 2025-01-08T10:45:46,299 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36283 {}] access.PermissionStorage(611): Read acl: entry[testExportWithResetTtl], kv [jenkins: RWXCA] 2025-01-08T10:45:46,301 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:45:46,301 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:45:46,301 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:45:46,301 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:45:46,304 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T10:45:46,304 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportWithResetTtl with data PBUF\x0A<\x0A\x07jenkins\x121\x08\x03"-\x0A!\x0A\x07default\x12\x16testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T10:45:46,304 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T10:45:46,305 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T10:45:46,305 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportWithResetTtl with data PBUF\x0A<\x0A\x07jenkins\x121\x08\x03"-\x0A!\x0A\x07default\x12\x16testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T10:45:46,305 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportWithResetTtl with data PBUF\x0A<\x0A\x07jenkins\x121\x08\x03"-\x0A!\x0A\x07default\x12\x16testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T10:45:46,305 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T10:45:46,305 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportWithResetTtl with data PBUF\x0A<\x0A\x07jenkins\x121\x08\x03"-\x0A!\x0A\x07default\x12\x16testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T10:45:46,306 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=41, state=SUCCESS; CreateTableProcedure table=testExportWithResetTtl in 815 msec 2025-01-08T10:45:46,598 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=41 2025-01-08T10:45:46,599 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testExportWithResetTtl, procId: 41 completed 2025-01-08T10:45:46,601 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testExportWithResetTtl 2025-01-08T10:45:46,601 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06. 2025-01-08T10:45:46,602 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T10:45:46,617 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=45003 {}] regionserver.HRegion(8254): writing data to region testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T10:45:46,620 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36283 {}] regionserver.HRegion(8254): writing data to region testExportWithResetTtl,1,1736333145488.d553a3cb3f214b19b4267c8eb74194a5. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T10:45:46,624 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testExportWithResetTtl 2025-01-08T10:45:46,624 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06. 2025-01-08T10:45:46,624 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T10:45:46,639 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.2 snapshot request for:{ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } 2025-01-08T10:45:46,639 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736333146639 (current time:1736333146639). 2025-01-08T10:45:46,640 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb-testExportWithResetTtl VERSION not specified, setting to 2 2025-01-08T10:45:46,640 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T10:45:46,641 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x2a72d85d to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@5f52b2dc 2025-01-08T10:45:46,646 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4a751a23, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:45:46,647 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:45:46,648 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:51022, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:45:46,649 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x2a72d85d to 127.0.0.1:52367 2025-01-08T10:45:46,650 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:45:46,651 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x0a0bb1fd to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@40053632 2025-01-08T10:45:46,654 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@22439813, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:45:46,656 DEBUG [hconnection-0x2d71163c-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:45:46,657 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:51034, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:45:46,660 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x0a0bb1fd to 127.0.0.1:52367 2025-01-08T10:45:46,660 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:45:46,660 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] access.PermissionStorage(611): Read acl: entry[testExportWithResetTtl], kv [jenkins: RWXCA] 2025-01-08T10:45:46,661 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T10:45:46,662 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=46, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } 2025-01-08T10:45:46,662 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 }, snapshot procedure id = 46 2025-01-08T10:45:46,663 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_PREPARE 2025-01-08T10:45:46,663 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=46 2025-01-08T10:45:46,664 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T10:45:46,666 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T10:45:46,678 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741924_1100 (size=143) 2025-01-08T10:45:46,678 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741924_1100 (size=143) 2025-01-08T10:45:46,678 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741924_1100 (size=143) 2025-01-08T10:45:46,679 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T10:45:46,680 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=47, ppid=46, state=RUNNABLE; SnapshotRegionProcedure 053a5459e246ced75c7eec0a1d77ff06}, {pid=48, ppid=46, state=RUNNABLE; SnapshotRegionProcedure d553a3cb3f214b19b4267c8eb74194a5}] 2025-01-08T10:45:46,681 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=48, ppid=46, state=RUNNABLE; SnapshotRegionProcedure d553a3cb3f214b19b4267c8eb74194a5 2025-01-08T10:45:46,681 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=47, ppid=46, state=RUNNABLE; SnapshotRegionProcedure 053a5459e246ced75c7eec0a1d77ff06 2025-01-08T10:45:46,764 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=46 2025-01-08T10:45:46,832 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:45:46,832 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:45:46,833 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=45003 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=47 2025-01-08T10:45:46,833 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=36283 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=48 2025-01-08T10:45:46,833 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testExportWithResetTtl,1,1736333145488.d553a3cb3f214b19b4267c8eb74194a5. 2025-01-08T10:45:46,833 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06. 2025-01-08T10:45:46,833 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.HRegion(2837): Flushing d553a3cb3f214b19b4267c8eb74194a5 1/1 column families, dataSize=3.06 KB heapSize=6.86 KB 2025-01-08T10:45:46,833 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.HRegion(2837): Flushing 053a5459e246ced75c7eec0a1d77ff06 1/1 column families, dataSize=199 B heapSize=688 B 2025-01-08T10:45:46,855 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108e386658c89a44383893b13b27e93a5c0_d553a3cb3f214b19b4267c8eb74194a5 is 71, key is 1020a5be276f62a986c37c3c38c06fd5/cf:q/1736333146619/Put/seqid=0 2025-01-08T10:45:46,866 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741925_1101 (size=8171) 2025-01-08T10:45:46,866 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741925_1101 (size=8171) 2025-01-08T10:45:46,867 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741925_1101 (size=8171) 2025-01-08T10:45:46,867 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:46,872 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108be6a03d9adc1435d92db5ba0211d5e56_053a5459e246ced75c7eec0a1d77ff06 is 71, key is 049d0d91ac0d0964e41c180a2040c818/cf:q/1736333146617/Put/seqid=0 2025-01-08T10:45:46,873 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108e386658c89a44383893b13b27e93a5c0_d553a3cb3f214b19b4267c8eb74194a5 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/c4ca4238a0b923820dcc509a6f75849b20250108e386658c89a44383893b13b27e93a5c0_d553a3cb3f214b19b4267c8eb74194a5 2025-01-08T10:45:46,875 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/d553a3cb3f214b19b4267c8eb74194a5/.tmp/cf/8841e72e7b1049d0b579bd16d3ea1517, store: [table=testExportWithResetTtl family=cf region=d553a3cb3f214b19b4267c8eb74194a5] 2025-01-08T10:45:46,875 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/d553a3cb3f214b19b4267c8eb74194a5/.tmp/cf/8841e72e7b1049d0b579bd16d3ea1517 is 199, key is 15f80aad30c5e37f0b21670292e4bb0d9/cf:q/1736333146619/Put/seqid=0 2025-01-08T10:45:46,882 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741926_1102 (size=5101) 2025-01-08T10:45:46,882 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741926_1102 (size=5101) 2025-01-08T10:45:46,882 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741926_1102 (size=5101) 2025-01-08T10:45:46,888 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741927_1103 (size=14517) 2025-01-08T10:45:46,889 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741927_1103 (size=14517) 2025-01-08T10:45:46,889 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741927_1103 (size=14517) 2025-01-08T10:45:46,890 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=5, memsize=3.1 K, hasBloomFilter=true, into tmp file hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/d553a3cb3f214b19b4267c8eb74194a5/.tmp/cf/8841e72e7b1049d0b579bd16d3ea1517 2025-01-08T10:45:46,898 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/d553a3cb3f214b19b4267c8eb74194a5/.tmp/cf/8841e72e7b1049d0b579bd16d3ea1517 as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/d553a3cb3f214b19b4267c8eb74194a5/cf/8841e72e7b1049d0b579bd16d3ea1517 2025-01-08T10:45:46,904 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/d553a3cb3f214b19b4267c8eb74194a5/cf/8841e72e7b1049d0b579bd16d3ea1517, entries=47, sequenceid=5, filesize=14.2 K 2025-01-08T10:45:46,906 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.HRegion(3040): Finished flush of dataSize ~3.06 KB/3137, heapSize ~6.84 KB/7008, currentSize=0 B/0 for d553a3cb3f214b19b4267c8eb74194a5 in 72ms, sequenceid=5, compaction requested=false 2025-01-08T10:45:46,906 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testExportWithResetTtl' 2025-01-08T10:45:46,907 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.HRegion(2538): Flush status journal for d553a3cb3f214b19b4267c8eb74194a5: 2025-01-08T10:45:46,907 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.SnapshotRegionCallable(75): Snapshotting region testExportWithResetTtl,1,1736333145488.d553a3cb3f214b19b4267c8eb74194a5. for snaptb-testExportWithResetTtl completed. 2025-01-08T10:45:46,907 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] snapshot.SnapshotManifest(241): Storing 'testExportWithResetTtl,1,1736333145488.d553a3cb3f214b19b4267c8eb74194a5.' region-info for snapshot=snaptb-testExportWithResetTtl 2025-01-08T10:45:46,907 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:45:46,907 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/d553a3cb3f214b19b4267c8eb74194a5/cf/8841e72e7b1049d0b579bd16d3ea1517] hfiles 2025-01-08T10:45:46,907 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/d553a3cb3f214b19b4267c8eb74194a5/cf/8841e72e7b1049d0b579bd16d3ea1517 for snapshot=snaptb-testExportWithResetTtl 2025-01-08T10:45:46,922 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741928_1104 (size=100) 2025-01-08T10:45:46,922 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741928_1104 (size=100) 2025-01-08T10:45:46,923 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741928_1104 (size=100) 2025-01-08T10:45:46,924 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testExportWithResetTtl,1,1736333145488.d553a3cb3f214b19b4267c8eb74194a5. 2025-01-08T10:45:46,924 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=48 2025-01-08T10:45:46,925 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=48 2025-01-08T10:45:46,925 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb-testExportWithResetTtl on region d553a3cb3f214b19b4267c8eb74194a5 2025-01-08T10:45:46,926 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=48, ppid=46, state=RUNNABLE; SnapshotRegionProcedure d553a3cb3f214b19b4267c8eb74194a5 2025-01-08T10:45:46,930 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=48, ppid=46, state=SUCCESS; SnapshotRegionProcedure d553a3cb3f214b19b4267c8eb74194a5 in 247 msec 2025-01-08T10:45:46,966 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=46 2025-01-08T10:45:47,267 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=46 2025-01-08T10:45:47,283 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:47,301 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108be6a03d9adc1435d92db5ba0211d5e56_053a5459e246ced75c7eec0a1d77ff06 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/d41d8cd98f00b204e9800998ecf8427e20250108be6a03d9adc1435d92db5ba0211d5e56_053a5459e246ced75c7eec0a1d77ff06 2025-01-08T10:45:47,303 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/053a5459e246ced75c7eec0a1d77ff06/.tmp/cf/583ae20d181b440d80e2fb8f69ac3773, store: [table=testExportWithResetTtl family=cf region=053a5459e246ced75c7eec0a1d77ff06] 2025-01-08T10:45:47,304 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/053a5459e246ced75c7eec0a1d77ff06/.tmp/cf/583ae20d181b440d80e2fb8f69ac3773 is 199, key is 0876b788f1cbe3acf95812a43c6cce662/cf:q/1736333146617/Put/seqid=0 2025-01-08T10:45:47,318 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741929_1105 (size=5876) 2025-01-08T10:45:47,318 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741929_1105 (size=5876) 2025-01-08T10:45:47,319 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741929_1105 (size=5876) 2025-01-08T10:45:47,319 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=5, memsize=199, hasBloomFilter=true, into tmp file hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/053a5459e246ced75c7eec0a1d77ff06/.tmp/cf/583ae20d181b440d80e2fb8f69ac3773 2025-01-08T10:45:47,328 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/053a5459e246ced75c7eec0a1d77ff06/.tmp/cf/583ae20d181b440d80e2fb8f69ac3773 as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/053a5459e246ced75c7eec0a1d77ff06/cf/583ae20d181b440d80e2fb8f69ac3773 2025-01-08T10:45:47,340 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/053a5459e246ced75c7eec0a1d77ff06/cf/583ae20d181b440d80e2fb8f69ac3773, entries=3, sequenceid=5, filesize=5.7 K 2025-01-08T10:45:47,343 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.HRegion(3040): Finished flush of dataSize ~199 B/199, heapSize ~672 B/672, currentSize=0 B/0 for 053a5459e246ced75c7eec0a1d77ff06 in 509ms, sequenceid=5, compaction requested=false 2025-01-08T10:45:47,343 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.HRegion(2538): Flush status journal for 053a5459e246ced75c7eec0a1d77ff06: 2025-01-08T10:45:47,343 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.SnapshotRegionCallable(75): Snapshotting region testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06. for snaptb-testExportWithResetTtl completed. 2025-01-08T10:45:47,343 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] snapshot.SnapshotManifest(241): Storing 'testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06.' region-info for snapshot=snaptb-testExportWithResetTtl 2025-01-08T10:45:47,343 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:45:47,343 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/053a5459e246ced75c7eec0a1d77ff06/cf/583ae20d181b440d80e2fb8f69ac3773] hfiles 2025-01-08T10:45:47,343 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/053a5459e246ced75c7eec0a1d77ff06/cf/583ae20d181b440d80e2fb8f69ac3773 for snapshot=snaptb-testExportWithResetTtl 2025-01-08T10:45:47,361 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741930_1106 (size=100) 2025-01-08T10:45:47,362 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741930_1106 (size=100) 2025-01-08T10:45:47,362 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741930_1106 (size=100) 2025-01-08T10:45:47,362 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06. 2025-01-08T10:45:47,363 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=47 2025-01-08T10:45:47,363 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=47 2025-01-08T10:45:47,363 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb-testExportWithResetTtl on region 053a5459e246ced75c7eec0a1d77ff06 2025-01-08T10:45:47,363 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=47, ppid=46, state=RUNNABLE; SnapshotRegionProcedure 053a5459e246ced75c7eec0a1d77ff06 2025-01-08T10:45:47,367 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=47, resume processing ppid=46 2025-01-08T10:45:47,367 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=47, ppid=46, state=SUCCESS; SnapshotRegionProcedure 053a5459e246ced75c7eec0a1d77ff06 in 685 msec 2025-01-08T10:45:47,367 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T10:45:47,368 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T10:45:47,370 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T10:45:47,370 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T10:45:47,370 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:45:47,374 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/c4ca4238a0b923820dcc509a6f75849b20250108e386658c89a44383893b13b27e93a5c0_d553a3cb3f214b19b4267c8eb74194a5, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/d41d8cd98f00b204e9800998ecf8427e20250108be6a03d9adc1435d92db5ba0211d5e56_053a5459e246ced75c7eec0a1d77ff06] hfiles 2025-01-08T10:45:47,374 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/c4ca4238a0b923820dcc509a6f75849b20250108e386658c89a44383893b13b27e93a5c0_d553a3cb3f214b19b4267c8eb74194a5 2025-01-08T10:45:47,374 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/d41d8cd98f00b204e9800998ecf8427e20250108be6a03d9adc1435d92db5ba0211d5e56_053a5459e246ced75c7eec0a1d77ff06 2025-01-08T10:45:47,399 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741931_1107 (size=284) 2025-01-08T10:45:47,400 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741931_1107 (size=284) 2025-01-08T10:45:47,400 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741931_1107 (size=284) 2025-01-08T10:45:47,406 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T10:45:47,406 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb-testExportWithResetTtl 2025-01-08T10:45:47,407 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/snaptb-testExportWithResetTtl 2025-01-08T10:45:47,424 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741932_1108 (size=923) 2025-01-08T10:45:47,427 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741932_1108 (size=923) 2025-01-08T10:45:47,427 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741932_1108 (size=923) 2025-01-08T10:45:47,433 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T10:45:47,448 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T10:45:47,449 DEBUG [PEWorker-3 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/snaptb-testExportWithResetTtl to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb-testExportWithResetTtl 2025-01-08T10:45:47,452 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T10:45:47,452 DEBUG [PEWorker-3 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 }, snapshot procedure id = 46 2025-01-08T10:45:47,454 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=46, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } in 791 msec 2025-01-08T10:45:47,502 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0001_000001 (auth:SIMPLE) from 127.0.0.1:38338 2025-01-08T10:45:47,512 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_0/usercache/jenkins/appcache/application_1736333116683_0001/container_1736333116683_0001_01_000001/launch_container.sh] 2025-01-08T10:45:47,512 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_0/usercache/jenkins/appcache/application_1736333116683_0001/container_1736333116683_0001_01_000001/container_tokens] 2025-01-08T10:45:47,512 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_0/usercache/jenkins/appcache/application_1736333116683_0001/container_1736333116683_0001_01_000001/sysfs] 2025-01-08T10:45:47,595 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testExportWithResetTtl 2025-01-08T10:45:47,596 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testExportWithResetTtl Metrics about Tables on a single HBase RegionServer 2025-01-08T10:45:47,597 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportWithResetTtl 2025-01-08T10:45:47,597 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportWithResetTtl Metrics about Tables on a single HBase RegionServer 2025-01-08T10:45:47,597 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportWithTargetName 2025-01-08T10:45:47,769 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=46 2025-01-08T10:45:47,770 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testExportWithResetTtl, procId: 46 completed 2025-01-08T10:45:47,780 INFO [Time-limited test {}] snapshot.TestExportSnapshot(468): HDFS export destination path: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333147780 2025-01-08T10:45:47,781 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=hdfs://localhost:37495, tgtDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333147780, rawTgtDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333147780, srcFsUri=hdfs://localhost:37495, srcDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:45:47,811 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:37495, inputRoot=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:45:47,811 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-832679797_22, ugi=jenkins (auth:SIMPLE)]], outputRoot=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333147780, skipTmp=false, initialOutputSnapshotDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333147780/.hbase-snapshot/.tmp/snaptb-testExportWithResetTtl 2025-01-08T10:45:47,813 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T10:45:47,818 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb-testExportWithResetTtl to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333147780/.hbase-snapshot/.tmp/snaptb-testExportWithResetTtl 2025-01-08T10:45:47,837 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741933_1109 (size=143) 2025-01-08T10:45:47,837 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741933_1109 (size=143) 2025-01-08T10:45:47,837 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741933_1109 (size=143) 2025-01-08T10:45:47,848 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741934_1110 (size=923) 2025-01-08T10:45:47,849 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741934_1110 (size=923) 2025-01-08T10:45:47,849 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741934_1110 (size=923) 2025-01-08T10:45:47,864 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741935_1111 (size=141) 2025-01-08T10:45:47,866 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741935_1111 (size=141) 2025-01-08T10:45:47,866 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741935_1111 (size=141) 2025-01-08T10:45:48,106 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-16481830410840616924.jar 2025-01-08T10:45:48,106 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T10:45:48,107 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T10:45:48,107 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T10:45:49,100 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T10:45:49,286 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-3678544516908395024.jar 2025-01-08T10:45:49,287 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T10:45:49,287 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T10:45:49,359 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-8473366533986095139.jar 2025-01-08T10:45:49,360 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T10:45:49,360 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T10:45:49,361 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T10:45:49,361 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T10:45:49,362 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T10:45:49,362 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T10:45:49,362 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T10:45:49,363 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T10:45:49,363 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T10:45:49,363 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T10:45:49,364 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T10:45:49,364 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T10:45:49,365 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T10:45:49,365 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T10:45:49,365 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T10:45:49,366 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T10:45:49,366 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T10:45:49,367 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T10:45:49,367 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:45:49,368 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:45:49,368 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:45:49,368 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:45:49,369 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:45:49,369 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:45:49,369 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:45:49,454 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741936_1112 (size=127628) 2025-01-08T10:45:49,455 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741936_1112 (size=127628) 2025-01-08T10:45:49,455 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741936_1112 (size=127628) 2025-01-08T10:45:49,472 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741937_1113 (size=2172137) 2025-01-08T10:45:49,473 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741937_1113 (size=2172137) 2025-01-08T10:45:49,473 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741937_1113 (size=2172137) 2025-01-08T10:45:49,484 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741938_1114 (size=213228) 2025-01-08T10:45:49,484 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741938_1114 (size=213228) 2025-01-08T10:45:49,485 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741938_1114 (size=213228) 2025-01-08T10:45:49,571 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741939_1115 (size=1877034) 2025-01-08T10:45:49,572 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741939_1115 (size=1877034) 2025-01-08T10:45:49,575 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741939_1115 (size=1877034) 2025-01-08T10:45:49,591 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741940_1116 (size=533455) 2025-01-08T10:45:49,591 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741940_1116 (size=533455) 2025-01-08T10:45:49,591 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741940_1116 (size=533455) 2025-01-08T10:45:49,621 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741941_1117 (size=7280644) 2025-01-08T10:45:49,621 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741941_1117 (size=7280644) 2025-01-08T10:45:49,622 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741941_1117 (size=7280644) 2025-01-08T10:45:49,665 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741942_1118 (size=4188619) 2025-01-08T10:45:49,665 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741942_1118 (size=4188619) 2025-01-08T10:45:49,666 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741942_1118 (size=4188619) 2025-01-08T10:45:49,684 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741943_1119 (size=20406) 2025-01-08T10:45:49,685 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741943_1119 (size=20406) 2025-01-08T10:45:49,685 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741943_1119 (size=20406) 2025-01-08T10:45:49,711 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741944_1120 (size=75495) 2025-01-08T10:45:49,712 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741944_1120 (size=75495) 2025-01-08T10:45:49,715 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741944_1120 (size=75495) 2025-01-08T10:45:49,749 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741945_1121 (size=45609) 2025-01-08T10:45:49,749 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741945_1121 (size=45609) 2025-01-08T10:45:49,751 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741945_1121 (size=45609) 2025-01-08T10:45:49,769 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741946_1122 (size=110084) 2025-01-08T10:45:49,769 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741946_1122 (size=110084) 2025-01-08T10:45:49,769 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741946_1122 (size=110084) 2025-01-08T10:45:49,822 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741947_1123 (size=1323991) 2025-01-08T10:45:49,823 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741947_1123 (size=1323991) 2025-01-08T10:45:49,824 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741947_1123 (size=1323991) 2025-01-08T10:45:49,865 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741948_1124 (size=23076) 2025-01-08T10:45:49,865 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741948_1124 (size=23076) 2025-01-08T10:45:49,866 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741948_1124 (size=23076) 2025-01-08T10:45:49,919 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741949_1125 (size=126803) 2025-01-08T10:45:49,919 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741949_1125 (size=126803) 2025-01-08T10:45:49,920 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741949_1125 (size=126803) 2025-01-08T10:45:49,950 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741950_1126 (size=322274) 2025-01-08T10:45:49,956 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741950_1126 (size=322274) 2025-01-08T10:45:49,964 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741950_1126 (size=322274) 2025-01-08T10:45:50,002 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741951_1127 (size=451756) 2025-01-08T10:45:50,003 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741951_1127 (size=451756) 2025-01-08T10:45:50,003 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741951_1127 (size=451756) 2025-01-08T10:45:50,035 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741952_1128 (size=1832290) 2025-01-08T10:45:50,035 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741952_1128 (size=1832290) 2025-01-08T10:45:50,035 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741952_1128 (size=1832290) 2025-01-08T10:45:50,068 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741953_1129 (size=912094) 2025-01-08T10:45:50,069 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741953_1129 (size=912094) 2025-01-08T10:45:50,069 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741953_1129 (size=912094) 2025-01-08T10:45:50,113 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741954_1130 (size=30081) 2025-01-08T10:45:50,113 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741954_1130 (size=30081) 2025-01-08T10:45:50,113 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741954_1130 (size=30081) 2025-01-08T10:45:50,161 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741955_1131 (size=53616) 2025-01-08T10:45:50,161 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741955_1131 (size=53616) 2025-01-08T10:45:50,165 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741955_1131 (size=53616) 2025-01-08T10:45:50,221 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741956_1132 (size=29229) 2025-01-08T10:45:50,221 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741956_1132 (size=29229) 2025-01-08T10:45:50,222 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741956_1132 (size=29229) 2025-01-08T10:45:50,266 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741957_1133 (size=169089) 2025-01-08T10:45:50,271 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741957_1133 (size=169089) 2025-01-08T10:45:50,272 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741957_1133 (size=169089) 2025-01-08T10:45:50,327 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741958_1134 (size=5175431) 2025-01-08T10:45:50,327 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741958_1134 (size=5175431) 2025-01-08T10:45:50,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741958_1134 (size=5175431) 2025-01-08T10:45:50,349 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741959_1135 (size=136454) 2025-01-08T10:45:50,350 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741959_1135 (size=136454) 2025-01-08T10:45:50,350 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741959_1135 (size=136454) 2025-01-08T10:45:50,411 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741960_1136 (size=6350714) 2025-01-08T10:45:50,414 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741960_1136 (size=6350714) 2025-01-08T10:45:50,414 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741960_1136 (size=6350714) 2025-01-08T10:45:50,448 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741961_1137 (size=3317408) 2025-01-08T10:45:50,450 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741961_1137 (size=3317408) 2025-01-08T10:45:50,457 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741961_1137 (size=3317408) 2025-01-08T10:45:50,482 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741962_1138 (size=503880) 2025-01-08T10:45:50,482 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741962_1138 (size=503880) 2025-01-08T10:45:50,482 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741962_1138 (size=503880) 2025-01-08T10:45:50,535 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741963_1139 (size=4695811) 2025-01-08T10:45:50,536 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741963_1139 (size=4695811) 2025-01-08T10:45:50,536 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741963_1139 (size=4695811) 2025-01-08T10:45:50,538 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T10:45:50,541 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'snaptb-testExportWithResetTtl' hfile list 2025-01-08T10:45:50,544 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(751): export split=0 size=32.9 K 2025-01-08T10:45:50,580 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741964_1140 (size=686) 2025-01-08T10:45:50,581 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741964_1140 (size=686) 2025-01-08T10:45:50,582 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741964_1140 (size=686) 2025-01-08T10:45:50,632 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741965_1141 (size=15) 2025-01-08T10:45:50,634 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741965_1141 (size=15) 2025-01-08T10:45:50,634 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741965_1141 (size=15) 2025-01-08T10:45:50,734 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741966_1142 (size=304923) 2025-01-08T10:45:50,734 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741966_1142 (size=304923) 2025-01-08T10:45:50,735 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741966_1142 (size=304923) 2025-01-08T10:45:50,762 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T10:45:50,762 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T10:45:51,415 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0002_000001 (auth:SIMPLE) from 127.0.0.1:41486 2025-01-08T10:45:52,460 INFO [master/67f42ee69b4e:0.Chore.1 {}] migrate.RollingUpgradeChore(116): There is no table to migrate StoreFileTracker! 2025-01-08T10:45:52,460 INFO [master/67f42ee69b4e:0.Chore.1 {}] migrate.RollingUpgradeChore(85): All Rolling-Upgrade tasks are complete, shutdown RollingUpgradeChore! 2025-01-08T10:45:57,866 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0002_000001 (auth:SIMPLE) from 127.0.0.1:56494 2025-01-08T10:45:58,240 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741967_1143 (size=350597) 2025-01-08T10:45:58,240 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741967_1143 (size=350597) 2025-01-08T10:45:58,241 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741967_1143 (size=350597) 2025-01-08T10:46:00,181 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0002_000001 (auth:SIMPLE) from 127.0.0.1:43718 2025-01-08T10:46:03,351 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741968_1144 (size=14517) 2025-01-08T10:46:03,351 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741968_1144 (size=14517) 2025-01-08T10:46:03,351 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741968_1144 (size=14517) 2025-01-08T10:46:03,416 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741969_1145 (size=8171) 2025-01-08T10:46:03,416 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741969_1145 (size=8171) 2025-01-08T10:46:03,416 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741969_1145 (size=8171) 2025-01-08T10:46:03,461 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741970_1146 (size=5876) 2025-01-08T10:46:03,462 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741970_1146 (size=5876) 2025-01-08T10:46:03,462 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741970_1146 (size=5876) 2025-01-08T10:46:03,529 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741971_1147 (size=5101) 2025-01-08T10:46:03,530 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741971_1147 (size=5101) 2025-01-08T10:46:03,530 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741971_1147 (size=5101) 2025-01-08T10:46:03,631 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741972_1148 (size=17458) 2025-01-08T10:46:03,631 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741972_1148 (size=17458) 2025-01-08T10:46:03,631 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741972_1148 (size=17458) 2025-01-08T10:46:03,658 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741973_1149 (size=461) 2025-01-08T10:46:03,658 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741973_1149 (size=461) 2025-01-08T10:46:03,658 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741973_1149 (size=461) 2025-01-08T10:46:03,698 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_2/usercache/jenkins/appcache/application_1736333116683_0002/container_1736333116683_0002_01_000002/launch_container.sh] 2025-01-08T10:46:03,698 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_2/usercache/jenkins/appcache/application_1736333116683_0002/container_1736333116683_0002_01_000002/container_tokens] 2025-01-08T10:46:03,698 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_2/usercache/jenkins/appcache/application_1736333116683_0002/container_1736333116683_0002_01_000002/sysfs] 2025-01-08T10:46:03,700 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741974_1150 (size=17458) 2025-01-08T10:46:03,701 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741974_1150 (size=17458) 2025-01-08T10:46:03,701 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741974_1150 (size=17458) 2025-01-08T10:46:03,723 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741975_1151 (size=350597) 2025-01-08T10:46:03,723 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741975_1151 (size=350597) 2025-01-08T10:46:03,723 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741975_1151 (size=350597) 2025-01-08T10:46:03,743 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0002_000001 (auth:SIMPLE) from 127.0.0.1:43728 2025-01-08T10:46:04,939 INFO [Time-limited test {}] snapshot.ExportSnapshot(1207): Finalize the Snapshot Export 2025-01-08T10:46:04,941 INFO [Time-limited test {}] snapshot.ExportSnapshot(1218): Verify the exported snapshot's expiration status and integrity. 2025-01-08T10:46:04,948 INFO [Time-limited test {}] snapshot.ExportSnapshot(1224): Export Completed: snaptb-testExportWithResetTtl 2025-01-08T10:46:04,949 INFO [Time-limited test {}] snapshot.TestExportSnapshot(362): Exported snapshot 2025-01-08T10:46:04,949 INFO [Time-limited test {}] snapshot.TestExportSnapshot(373): Verified filesystem state 2025-01-08T10:46:04,949 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-832679797_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb-testExportWithResetTtl at hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb-testExportWithResetTtl 2025-01-08T10:46:04,950 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb-testExportWithResetTtl/.snapshotinfo 2025-01-08T10:46:04,950 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb-testExportWithResetTtl/data.manifest 2025-01-08T10:46:04,950 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-832679797_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333147780/.hbase-snapshot/snaptb-testExportWithResetTtl at hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333147780/.hbase-snapshot/snaptb-testExportWithResetTtl 2025-01-08T10:46:04,950 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333147780/.hbase-snapshot/snaptb-testExportWithResetTtl/.snapshotinfo 2025-01-08T10:46:04,950 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333147780/.hbase-snapshot/snaptb-testExportWithResetTtl/data.manifest 2025-01-08T10:46:04,958 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testExportWithResetTtl 2025-01-08T10:46:04,958 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.2 disable testExportWithResetTtl 2025-01-08T10:46:04,959 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=49, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testExportWithResetTtl 2025-01-08T10:46:04,961 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=49 2025-01-08T10:46:04,961 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333164961"}]},"ts":"1736333164961"} 2025-01-08T10:46:04,963 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testExportWithResetTtl, state=DISABLING in hbase:meta 2025-01-08T10:46:04,965 INFO [PEWorker-4 {}] procedure.DisableTableProcedure(284): Set testExportWithResetTtl to state=DISABLING 2025-01-08T10:46:04,966 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=50, ppid=49, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testExportWithResetTtl}] 2025-01-08T10:46:04,967 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=51, ppid=50, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testExportWithResetTtl, region=053a5459e246ced75c7eec0a1d77ff06, UNASSIGN}, {pid=52, ppid=50, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testExportWithResetTtl, region=d553a3cb3f214b19b4267c8eb74194a5, UNASSIGN}] 2025-01-08T10:46:04,968 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=52, ppid=50, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testExportWithResetTtl, region=d553a3cb3f214b19b4267c8eb74194a5, UNASSIGN 2025-01-08T10:46:04,968 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=51, ppid=50, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testExportWithResetTtl, region=053a5459e246ced75c7eec0a1d77ff06, UNASSIGN 2025-01-08T10:46:04,969 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=51 updating hbase:meta row=053a5459e246ced75c7eec0a1d77ff06, regionState=CLOSING, regionLocation=67f42ee69b4e,45003,1736333108039 2025-01-08T10:46:04,969 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=52 updating hbase:meta row=d553a3cb3f214b19b4267c8eb74194a5, regionState=CLOSING, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:46:04,971 DEBUG [PEWorker-2 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T10:46:04,971 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=53, ppid=51, state=RUNNABLE; CloseRegionProcedure 053a5459e246ced75c7eec0a1d77ff06, server=67f42ee69b4e,45003,1736333108039}] 2025-01-08T10:46:04,972 DEBUG [PEWorker-5 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T10:46:04,972 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=54, ppid=52, state=RUNNABLE; CloseRegionProcedure d553a3cb3f214b19b4267c8eb74194a5, server=67f42ee69b4e,36283,1736333108166}] 2025-01-08T10:46:05,062 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=49 2025-01-08T10:46:05,123 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:46:05,124 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] handler.UnassignRegionHandler(124): Close 053a5459e246ced75c7eec0a1d77ff06 2025-01-08T10:46:05,124 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:46:05,124 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T10:46:05,124 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] regionserver.HRegion(1681): Closing 053a5459e246ced75c7eec0a1d77ff06, disabling compactions & flushes 2025-01-08T10:46:05,124 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] regionserver.HRegion(1703): Closing region testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06. 2025-01-08T10:46:05,124 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] handler.UnassignRegionHandler(124): Close d553a3cb3f214b19b4267c8eb74194a5 2025-01-08T10:46:05,124 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] regionserver.HRegion(1724): Waiting without time limit for close lock on testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06. 2025-01-08T10:46:05,124 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] regionserver.HRegion(1791): Acquired close lock on testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06. after waiting 0 ms 2025-01-08T10:46:05,124 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] regionserver.HRegion(1801): Updates disabled for region testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06. 2025-01-08T10:46:05,124 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T10:46:05,124 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] regionserver.HRegion(1681): Closing d553a3cb3f214b19b4267c8eb74194a5, disabling compactions & flushes 2025-01-08T10:46:05,124 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] regionserver.HRegion(1703): Closing region testExportWithResetTtl,1,1736333145488.d553a3cb3f214b19b4267c8eb74194a5. 2025-01-08T10:46:05,124 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] regionserver.HRegion(1724): Waiting without time limit for close lock on testExportWithResetTtl,1,1736333145488.d553a3cb3f214b19b4267c8eb74194a5. 2025-01-08T10:46:05,124 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] regionserver.HRegion(1791): Acquired close lock on testExportWithResetTtl,1,1736333145488.d553a3cb3f214b19b4267c8eb74194a5. after waiting 0 ms 2025-01-08T10:46:05,124 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] regionserver.HRegion(1801): Updates disabled for region testExportWithResetTtl,1,1736333145488.d553a3cb3f214b19b4267c8eb74194a5. 2025-01-08T10:46:05,130 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/053a5459e246ced75c7eec0a1d77ff06/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2025-01-08T10:46:05,131 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/d553a3cb3f214b19b4267c8eb74194a5/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2025-01-08T10:46:05,131 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:46:05,131 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:46:05,131 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] regionserver.HRegion(1922): Closed testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06. 2025-01-08T10:46:05,131 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] regionserver.HRegion(1922): Closed testExportWithResetTtl,1,1736333145488.d553a3cb3f214b19b4267c8eb74194a5. 2025-01-08T10:46:05,131 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] regionserver.HRegion(1635): Region close journal for 053a5459e246ced75c7eec0a1d77ff06: 2025-01-08T10:46:05,131 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] regionserver.HRegion(1635): Region close journal for d553a3cb3f214b19b4267c8eb74194a5: 2025-01-08T10:46:05,133 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] handler.UnassignRegionHandler(170): Closed 053a5459e246ced75c7eec0a1d77ff06 2025-01-08T10:46:05,134 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=51 updating hbase:meta row=053a5459e246ced75c7eec0a1d77ff06, regionState=CLOSED 2025-01-08T10:46:05,134 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] handler.UnassignRegionHandler(170): Closed d553a3cb3f214b19b4267c8eb74194a5 2025-01-08T10:46:05,134 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=52 updating hbase:meta row=d553a3cb3f214b19b4267c8eb74194a5, regionState=CLOSED 2025-01-08T10:46:05,137 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=53, resume processing ppid=51 2025-01-08T10:46:05,142 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=53, ppid=51, state=SUCCESS; CloseRegionProcedure 053a5459e246ced75c7eec0a1d77ff06, server=67f42ee69b4e,45003,1736333108039 in 164 msec 2025-01-08T10:46:05,143 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=54, resume processing ppid=52 2025-01-08T10:46:05,143 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=51, ppid=50, state=SUCCESS; TransitRegionStateProcedure table=testExportWithResetTtl, region=053a5459e246ced75c7eec0a1d77ff06, UNASSIGN in 170 msec 2025-01-08T10:46:05,143 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=54, ppid=52, state=SUCCESS; CloseRegionProcedure d553a3cb3f214b19b4267c8eb74194a5, server=67f42ee69b4e,36283,1736333108166 in 164 msec 2025-01-08T10:46:05,145 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=52, resume processing ppid=50 2025-01-08T10:46:05,145 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=52, ppid=50, state=SUCCESS; TransitRegionStateProcedure table=testExportWithResetTtl, region=d553a3cb3f214b19b4267c8eb74194a5, UNASSIGN in 176 msec 2025-01-08T10:46:05,147 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=50, resume processing ppid=49 2025-01-08T10:46:05,147 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=50, ppid=49, state=SUCCESS; CloseTableRegionsProcedure table=testExportWithResetTtl in 179 msec 2025-01-08T10:46:05,148 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333165148"}]},"ts":"1736333165148"} 2025-01-08T10:46:05,150 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1655): Updated tableName=testExportWithResetTtl, state=DISABLED in hbase:meta 2025-01-08T10:46:05,151 INFO [PEWorker-1 {}] procedure.DisableTableProcedure(296): Set testExportWithResetTtl to state=DISABLED 2025-01-08T10:46:05,153 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=49, state=SUCCESS; DisableTableProcedure table=testExportWithResetTtl in 194 msec 2025-01-08T10:46:05,264 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=49 2025-01-08T10:46:05,265 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testExportWithResetTtl, procId: 49 completed 2025-01-08T10:46:05,266 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.2 delete testExportWithResetTtl 2025-01-08T10:46:05,267 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=55, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testExportWithResetTtl 2025-01-08T10:46:05,269 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=55, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testExportWithResetTtl 2025-01-08T10:46:05,269 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] access.PermissionStorage(259): Removing permissions of removed table testExportWithResetTtl 2025-01-08T10:46:05,270 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=55, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testExportWithResetTtl 2025-01-08T10:46:05,272 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36283 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testExportWithResetTtl 2025-01-08T10:46:05,275 DEBUG [HFileArchiver-3 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/053a5459e246ced75c7eec0a1d77ff06 2025-01-08T10:46:05,275 DEBUG [HFileArchiver-4 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/d553a3cb3f214b19b4267c8eb74194a5 2025-01-08T10:46:05,275 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testExportWithResetTtl 2025-01-08T10:46:05,275 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testExportWithResetTtl 2025-01-08T10:46:05,275 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testExportWithResetTtl 2025-01-08T10:46:05,276 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testExportWithResetTtl 2025-01-08T10:46:05,276 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportWithResetTtl with data PBUF 2025-01-08T10:46:05,276 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportWithResetTtl with data PBUF 2025-01-08T10:46:05,277 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportWithResetTtl with data PBUF 2025-01-08T10:46:05,278 DEBUG [HFileArchiver-3 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/053a5459e246ced75c7eec0a1d77ff06/cf, FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/053a5459e246ced75c7eec0a1d77ff06/recovered.edits] 2025-01-08T10:46:05,278 DEBUG [HFileArchiver-4 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/d553a3cb3f214b19b4267c8eb74194a5/cf, FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/d553a3cb3f214b19b4267c8eb74194a5/recovered.edits] 2025-01-08T10:46:05,282 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportWithResetTtl with data PBUF 2025-01-08T10:46:05,282 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testExportWithResetTtl 2025-01-08T10:46:05,282 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testExportWithResetTtl 2025-01-08T10:46:05,282 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testExportWithResetTtl 2025-01-08T10:46:05,282 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:46:05,282 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:46:05,282 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:46:05,283 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testExportWithResetTtl 2025-01-08T10:46:05,283 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:46:05,284 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T10:46:05,284 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T10:46:05,285 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T10:46:05,285 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=55 2025-01-08T10:46:05,285 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T10:46:05,286 DEBUG [HFileArchiver-3 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/053a5459e246ced75c7eec0a1d77ff06/cf/583ae20d181b440d80e2fb8f69ac3773 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testExportWithResetTtl/053a5459e246ced75c7eec0a1d77ff06/cf/583ae20d181b440d80e2fb8f69ac3773 2025-01-08T10:46:05,286 DEBUG [HFileArchiver-4 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/d553a3cb3f214b19b4267c8eb74194a5/cf/8841e72e7b1049d0b579bd16d3ea1517 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testExportWithResetTtl/d553a3cb3f214b19b4267c8eb74194a5/cf/8841e72e7b1049d0b579bd16d3ea1517 2025-01-08T10:46:05,290 DEBUG [HFileArchiver-4 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/d553a3cb3f214b19b4267c8eb74194a5/recovered.edits/8.seqid to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testExportWithResetTtl/d553a3cb3f214b19b4267c8eb74194a5/recovered.edits/8.seqid 2025-01-08T10:46:05,290 DEBUG [HFileArchiver-3 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/053a5459e246ced75c7eec0a1d77ff06/recovered.edits/8.seqid to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testExportWithResetTtl/053a5459e246ced75c7eec0a1d77ff06/recovered.edits/8.seqid 2025-01-08T10:46:05,291 DEBUG [HFileArchiver-4 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/d553a3cb3f214b19b4267c8eb74194a5 2025-01-08T10:46:05,291 DEBUG [HFileArchiver-3 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportWithResetTtl/053a5459e246ced75c7eec0a1d77ff06 2025-01-08T10:46:05,291 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(313): Archived testExportWithResetTtl regions 2025-01-08T10:46:05,292 DEBUG [PEWorker-5 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d 2025-01-08T10:46:05,293 DEBUG [PEWorker-5 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf] 2025-01-08T10:46:05,298 DEBUG [PEWorker-5 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/c4ca4238a0b923820dcc509a6f75849b20250108e386658c89a44383893b13b27e93a5c0_d553a3cb3f214b19b4267c8eb74194a5 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/c4ca4238a0b923820dcc509a6f75849b20250108e386658c89a44383893b13b27e93a5c0_d553a3cb3f214b19b4267c8eb74194a5 2025-01-08T10:46:05,300 DEBUG [PEWorker-5 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/d41d8cd98f00b204e9800998ecf8427e20250108be6a03d9adc1435d92db5ba0211d5e56_053a5459e246ced75c7eec0a1d77ff06 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/d41d8cd98f00b204e9800998ecf8427e20250108be6a03d9adc1435d92db5ba0211d5e56_053a5459e246ced75c7eec0a1d77ff06 2025-01-08T10:46:05,301 DEBUG [PEWorker-5 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d 2025-01-08T10:46:05,304 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=55, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testExportWithResetTtl 2025-01-08T10:46:05,307 WARN [PEWorker-5 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testExportWithResetTtl from hbase:meta 2025-01-08T10:46:05,310 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(407): Removing 'testExportWithResetTtl' descriptor. 2025-01-08T10:46:05,312 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=55, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testExportWithResetTtl 2025-01-08T10:46:05,313 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(397): Removing 'testExportWithResetTtl' from region states. 2025-01-08T10:46:05,313 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736333165313"}]},"ts":"9223372036854775807"} 2025-01-08T10:46:05,313 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testExportWithResetTtl,1,1736333145488.d553a3cb3f214b19b4267c8eb74194a5.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736333165313"}]},"ts":"9223372036854775807"} 2025-01-08T10:46:05,317 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T10:46:05,317 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => 053a5459e246ced75c7eec0a1d77ff06, NAME => 'testExportWithResetTtl,,1736333145488.053a5459e246ced75c7eec0a1d77ff06.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => d553a3cb3f214b19b4267c8eb74194a5, NAME => 'testExportWithResetTtl,1,1736333145488.d553a3cb3f214b19b4267c8eb74194a5.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T10:46:05,317 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(401): Marking 'testExportWithResetTtl' as deleted. 2025-01-08T10:46:05,318 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736333165317"}]},"ts":"9223372036854775807"} 2025-01-08T10:46:05,321 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1678): Deleted table testExportWithResetTtl state from META 2025-01-08T10:46:05,324 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(133): Finished pid=55, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testExportWithResetTtl 2025-01-08T10:46:05,325 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=55, state=SUCCESS; DeleteTableProcedure table=testExportWithResetTtl in 58 msec 2025-01-08T10:46:05,386 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=55 2025-01-08T10:46:05,387 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testExportWithResetTtl, procId: 55 completed 2025-01-08T10:46:05,387 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testExportWithResetTtl 2025-01-08T10:46:05,388 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.2 disable testtb-testExportWithResetTtl 2025-01-08T10:46:05,389 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=56, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testExportWithResetTtl 2025-01-08T10:46:05,394 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=56 2025-01-08T10:46:05,395 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333165395"}]},"ts":"1736333165395"} 2025-01-08T10:46:05,397 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithResetTtl, state=DISABLING in hbase:meta 2025-01-08T10:46:05,400 INFO [PEWorker-2 {}] procedure.DisableTableProcedure(284): Set testtb-testExportWithResetTtl to state=DISABLING 2025-01-08T10:46:05,401 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=57, ppid=56, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testExportWithResetTtl}] 2025-01-08T10:46:05,403 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=58, ppid=57, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=af16492a38d4fc130200d4b3697ff170, UNASSIGN}, {pid=59, ppid=57, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=45518da4f8eaa6bc80e04f2fcdc6fc93, UNASSIGN}] 2025-01-08T10:46:05,404 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=59, ppid=57, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=45518da4f8eaa6bc80e04f2fcdc6fc93, UNASSIGN 2025-01-08T10:46:05,404 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=58, ppid=57, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=af16492a38d4fc130200d4b3697ff170, UNASSIGN 2025-01-08T10:46:05,405 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=58 updating hbase:meta row=af16492a38d4fc130200d4b3697ff170, regionState=CLOSING, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:46:05,405 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=59 updating hbase:meta row=45518da4f8eaa6bc80e04f2fcdc6fc93, regionState=CLOSING, regionLocation=67f42ee69b4e,45003,1736333108039 2025-01-08T10:46:05,408 DEBUG [PEWorker-4 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T10:46:05,408 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=60, ppid=59, state=RUNNABLE; CloseRegionProcedure 45518da4f8eaa6bc80e04f2fcdc6fc93, server=67f42ee69b4e,45003,1736333108039}] 2025-01-08T10:46:05,409 DEBUG [PEWorker-1 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T10:46:05,409 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=61, ppid=58, state=RUNNABLE; CloseRegionProcedure af16492a38d4fc130200d4b3697ff170, server=67f42ee69b4e,36283,1736333108166}] 2025-01-08T10:46:05,495 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=56 2025-01-08T10:46:05,560 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:46:05,561 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] handler.UnassignRegionHandler(124): Close 45518da4f8eaa6bc80e04f2fcdc6fc93 2025-01-08T10:46:05,561 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T10:46:05,561 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] regionserver.HRegion(1681): Closing 45518da4f8eaa6bc80e04f2fcdc6fc93, disabling compactions & flushes 2025-01-08T10:46:05,561 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:46:05,561 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] regionserver.HRegion(1703): Closing region testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93. 2025-01-08T10:46:05,562 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93. 2025-01-08T10:46:05,562 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93. after waiting 0 ms 2025-01-08T10:46:05,562 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93. 2025-01-08T10:46:05,562 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] handler.UnassignRegionHandler(124): Close af16492a38d4fc130200d4b3697ff170 2025-01-08T10:46:05,562 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T10:46:05,562 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] regionserver.HRegion(1681): Closing af16492a38d4fc130200d4b3697ff170, disabling compactions & flushes 2025-01-08T10:46:05,562 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] regionserver.HRegion(1703): Closing region testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170. 2025-01-08T10:46:05,562 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170. 2025-01-08T10:46:05,562 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170. after waiting 0 ms 2025-01-08T10:46:05,562 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170. 2025-01-08T10:46:05,567 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/af16492a38d4fc130200d4b3697ff170/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T10:46:05,567 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:46:05,568 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/45518da4f8eaa6bc80e04f2fcdc6fc93/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T10:46:05,568 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] regionserver.HRegion(1922): Closed testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170. 2025-01-08T10:46:05,568 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] regionserver.HRegion(1635): Region close journal for af16492a38d4fc130200d4b3697ff170: 2025-01-08T10:46:05,568 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:46:05,568 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] regionserver.HRegion(1922): Closed testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93. 2025-01-08T10:46:05,568 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] regionserver.HRegion(1635): Region close journal for 45518da4f8eaa6bc80e04f2fcdc6fc93: 2025-01-08T10:46:05,569 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] handler.UnassignRegionHandler(170): Closed af16492a38d4fc130200d4b3697ff170 2025-01-08T10:46:05,570 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=58 updating hbase:meta row=af16492a38d4fc130200d4b3697ff170, regionState=CLOSED 2025-01-08T10:46:05,570 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] handler.UnassignRegionHandler(170): Closed 45518da4f8eaa6bc80e04f2fcdc6fc93 2025-01-08T10:46:05,574 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=59 updating hbase:meta row=45518da4f8eaa6bc80e04f2fcdc6fc93, regionState=CLOSED 2025-01-08T10:46:05,576 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=61, resume processing ppid=58 2025-01-08T10:46:05,576 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=61, ppid=58, state=SUCCESS; CloseRegionProcedure af16492a38d4fc130200d4b3697ff170, server=67f42ee69b4e,36283,1736333108166 in 163 msec 2025-01-08T10:46:05,577 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=60, resume processing ppid=59 2025-01-08T10:46:05,577 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=58, ppid=57, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=af16492a38d4fc130200d4b3697ff170, UNASSIGN in 173 msec 2025-01-08T10:46:05,577 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=60, ppid=59, state=SUCCESS; CloseRegionProcedure 45518da4f8eaa6bc80e04f2fcdc6fc93, server=67f42ee69b4e,45003,1736333108039 in 168 msec 2025-01-08T10:46:05,579 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=59, resume processing ppid=57 2025-01-08T10:46:05,579 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=59, ppid=57, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=45518da4f8eaa6bc80e04f2fcdc6fc93, UNASSIGN in 174 msec 2025-01-08T10:46:05,581 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=57, resume processing ppid=56 2025-01-08T10:46:05,581 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=57, ppid=56, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testExportWithResetTtl in 179 msec 2025-01-08T10:46:05,582 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333165582"}]},"ts":"1736333165582"} 2025-01-08T10:46:05,584 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithResetTtl, state=DISABLED in hbase:meta 2025-01-08T10:46:05,586 INFO [PEWorker-3 {}] procedure.DisableTableProcedure(296): Set testtb-testExportWithResetTtl to state=DISABLED 2025-01-08T10:46:05,587 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=56, state=SUCCESS; DisableTableProcedure table=testtb-testExportWithResetTtl in 198 msec 2025-01-08T10:46:05,697 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=56 2025-01-08T10:46:05,697 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testExportWithResetTtl, procId: 56 completed 2025-01-08T10:46:05,698 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.2 delete testtb-testExportWithResetTtl 2025-01-08T10:46:05,699 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=62, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testExportWithResetTtl 2025-01-08T10:46:05,700 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=62, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportWithResetTtl 2025-01-08T10:46:05,700 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testExportWithResetTtl 2025-01-08T10:46:05,701 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=62, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testExportWithResetTtl 2025-01-08T10:46:05,702 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36283 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testExportWithResetTtl 2025-01-08T10:46:05,706 DEBUG [HFileArchiver-5 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/af16492a38d4fc130200d4b3697ff170 2025-01-08T10:46:05,706 DEBUG [HFileArchiver-6 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/45518da4f8eaa6bc80e04f2fcdc6fc93 2025-01-08T10:46:05,706 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithResetTtl 2025-01-08T10:46:05,706 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithResetTtl 2025-01-08T10:46:05,707 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithResetTtl 2025-01-08T10:46:05,708 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithResetTtl 2025-01-08T10:46:05,709 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF 2025-01-08T10:46:05,709 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF 2025-01-08T10:46:05,709 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF 2025-01-08T10:46:05,709 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF 2025-01-08T10:46:05,709 DEBUG [HFileArchiver-6 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/45518da4f8eaa6bc80e04f2fcdc6fc93/cf, FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/45518da4f8eaa6bc80e04f2fcdc6fc93/recovered.edits] 2025-01-08T10:46:05,709 DEBUG [HFileArchiver-5 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/af16492a38d4fc130200d4b3697ff170/cf, FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/af16492a38d4fc130200d4b3697ff170/recovered.edits] 2025-01-08T10:46:05,711 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportWithResetTtl 2025-01-08T10:46:05,711 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportWithResetTtl 2025-01-08T10:46:05,711 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:46:05,711 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:46:05,711 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportWithResetTtl 2025-01-08T10:46:05,711 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportWithResetTtl 2025-01-08T10:46:05,711 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:46:05,712 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:46:05,713 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=62 2025-01-08T10:46:05,713 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T10:46:05,716 DEBUG [HFileArchiver-5 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/af16492a38d4fc130200d4b3697ff170/cf/0ea6716cd7184a7f9071648966df2e6c to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportWithResetTtl/af16492a38d4fc130200d4b3697ff170/cf/0ea6716cd7184a7f9071648966df2e6c 2025-01-08T10:46:05,716 DEBUG [HFileArchiver-6 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/45518da4f8eaa6bc80e04f2fcdc6fc93/cf/d4088f10677346b5b0f9e22c32824a43 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportWithResetTtl/45518da4f8eaa6bc80e04f2fcdc6fc93/cf/d4088f10677346b5b0f9e22c32824a43 2025-01-08T10:46:05,721 DEBUG [HFileArchiver-6 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/45518da4f8eaa6bc80e04f2fcdc6fc93/recovered.edits/9.seqid to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportWithResetTtl/45518da4f8eaa6bc80e04f2fcdc6fc93/recovered.edits/9.seqid 2025-01-08T10:46:05,722 DEBUG [HFileArchiver-5 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/af16492a38d4fc130200d4b3697ff170/recovered.edits/9.seqid to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportWithResetTtl/af16492a38d4fc130200d4b3697ff170/recovered.edits/9.seqid 2025-01-08T10:46:05,722 DEBUG [HFileArchiver-6 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/45518da4f8eaa6bc80e04f2fcdc6fc93 2025-01-08T10:46:05,722 DEBUG [HFileArchiver-5 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithResetTtl/af16492a38d4fc130200d4b3697ff170 2025-01-08T10:46:05,722 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(313): Archived testtb-testExportWithResetTtl regions 2025-01-08T10:46:05,723 DEBUG [PEWorker-1 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e 2025-01-08T10:46:05,724 DEBUG [PEWorker-1 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf] 2025-01-08T10:46:05,728 DEBUG [PEWorker-1 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/c4ca4238a0b923820dcc509a6f75849b20250108089b7e8178ce411e861aa3c3847e56cf_45518da4f8eaa6bc80e04f2fcdc6fc93 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/c4ca4238a0b923820dcc509a6f75849b20250108089b7e8178ce411e861aa3c3847e56cf_45518da4f8eaa6bc80e04f2fcdc6fc93 2025-01-08T10:46:05,730 DEBUG [PEWorker-1 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/d41d8cd98f00b204e9800998ecf8427e20250108870de8e9e6324e44bef2c74945990200_af16492a38d4fc130200d4b3697ff170 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/d41d8cd98f00b204e9800998ecf8427e20250108870de8e9e6324e44bef2c74945990200_af16492a38d4fc130200d4b3697ff170 2025-01-08T10:46:05,731 DEBUG [PEWorker-1 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e 2025-01-08T10:46:05,733 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=62, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testExportWithResetTtl 2025-01-08T10:46:05,736 WARN [PEWorker-1 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testtb-testExportWithResetTtl from hbase:meta 2025-01-08T10:46:05,738 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testExportWithResetTtl' descriptor. 2025-01-08T10:46:05,739 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=62, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testExportWithResetTtl 2025-01-08T10:46:05,740 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testExportWithResetTtl' from region states. 2025-01-08T10:46:05,740 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736333165740"}]},"ts":"9223372036854775807"} 2025-01-08T10:46:05,740 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736333165740"}]},"ts":"9223372036854775807"} 2025-01-08T10:46:05,742 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T10:46:05,742 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => af16492a38d4fc130200d4b3697ff170, NAME => 'testtb-testExportWithResetTtl,,1736333143867.af16492a38d4fc130200d4b3697ff170.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => 45518da4f8eaa6bc80e04f2fcdc6fc93, NAME => 'testtb-testExportWithResetTtl,1,1736333143867.45518da4f8eaa6bc80e04f2fcdc6fc93.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T10:46:05,742 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testExportWithResetTtl' as deleted. 2025-01-08T10:46:05,742 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736333165742"}]},"ts":"9223372036854775807"} 2025-01-08T10:46:05,744 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testExportWithResetTtl state from META 2025-01-08T10:46:05,746 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(133): Finished pid=62, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportWithResetTtl 2025-01-08T10:46:05,748 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=62, state=SUCCESS; DeleteTableProcedure table=testtb-testExportWithResetTtl in 48 msec 2025-01-08T10:46:05,814 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=62 2025-01-08T10:46:05,815 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testExportWithResetTtl, procId: 62 completed 2025-01-08T10:46:05,832 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.2 delete name: "emptySnaptb0-testExportWithResetTtl" 2025-01-08T10:46:05,835 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(380): Deleting snapshot: emptySnaptb0-testExportWithResetTtl 2025-01-08T10:46:05,837 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.2 delete name: "snaptb-testExportWithResetTtl" 2025-01-08T10:46:05,840 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb-testExportWithResetTtl 2025-01-08T10:46:05,842 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.2 delete name: "snaptb0-testExportWithResetTtl" 2025-01-08T10:46:05,844 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testExportWithResetTtl 2025-01-08T10:46:05,872 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: snapshot.TestMobSecureExportSnapshot#testExportWithResetTtl Thread=778 (was 771) Potentially hanging thread: ApplicationMasterLauncher #2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_388043009_1 at /127.0.0.1:57728 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-13 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_388043009_1 at /127.0.0.1:58602 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ContainersLauncher #1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:401) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-5 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-16 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-11 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-14 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-832679797_22 at /127.0.0.1:46266 [Waiting for operation #4] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-6 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-12 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LogDeleter #0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-15 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (175901922) connection to localhost/127.0.0.1:33775 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: HFileArchiver-4 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-832679797_22 at /127.0.0.1:43612 [Waiting for operation #6] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:45269 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-3 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:40861 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-832679797_22 at /127.0.0.1:40310 [Waiting for operation #4] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Thread-2251 java.base@17.0.11/java.io.FileInputStream.readBytes(Native Method) java.base@17.0.11/java.io.FileInputStream.read(FileInputStream.java:276) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:282) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) java.base@17.0.11/sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:281) java.base@17.0.11/sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:324) java.base@17.0.11/sun.nio.cs.StreamDecoder.read(StreamDecoder.java:189) java.base@17.0.11/java.io.InputStreamReader.read(InputStreamReader.java:177) java.base@17.0.11/java.io.BufferedReader.fill(BufferedReader.java:162) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:329) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:396) app//org.apache.hadoop.util.Shell$1.run(Shell.java:1025) Potentially hanging thread: process reaper (pid 9881) java.base@17.0.11/java.lang.ProcessHandleImpl.waitForProcessExit0(Native Method) java.base@17.0.11/java.lang.ProcessHandleImpl$1.run(ProcessHandleImpl.java:150) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LogDeleter #0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:33775 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) - Thread LEAK? -, OpenFileDescriptor=805 (was 813), MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=710 (was 802), ProcessCount=18 (was 20), AvailableMemoryMB=2137 (was 1210) - AvailableMemoryMB LEAK? - 2025-01-08T10:46:05,872 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=778 is superior to 500 2025-01-08T10:46:05,889 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: snapshot.TestMobSecureExportSnapshot#testExportFileSystemState Thread=778, OpenFileDescriptor=805, MaxFileDescriptor=1048576, SystemLoadAverage=710, ProcessCount=18, AvailableMemoryMB=2137 2025-01-08T10:46:05,889 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=778 is superior to 500 2025-01-08T10:46:05,891 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.2 create 'testtb-testExportFileSystemState', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T10:46:05,892 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=63, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testExportFileSystemState 2025-01-08T10:46:05,893 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=63, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportFileSystemState execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T10:46:05,893 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "testtb-testExportFileSystemState" procId is: 63 2025-01-08T10:46:05,894 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=63, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testExportFileSystemState execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T10:46:05,894 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=63 2025-01-08T10:46:05,907 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741976_1152 (size=443) 2025-01-08T10:46:05,907 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741976_1152 (size=443) 2025-01-08T10:46:05,907 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741976_1152 (size=443) 2025-01-08T10:46:05,913 INFO [RegionOpenAndInit-testtb-testExportFileSystemState-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => f2cb5b6ce155a147a1798fb975fd564d, NAME => 'testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testtb-testExportFileSystemState', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:46:05,913 INFO [RegionOpenAndInit-testtb-testExportFileSystemState-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 9f13bcd219cfe423da66381b35d7e9b8, NAME => 'testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testtb-testExportFileSystemState', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:46:05,924 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741978_1154 (size=68) 2025-01-08T10:46:05,925 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741978_1154 (size=68) 2025-01-08T10:46:05,925 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741977_1153 (size=68) 2025-01-08T10:46:05,925 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741977_1153 (size=68) 2025-01-08T10:46:05,926 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741978_1154 (size=68) 2025-01-08T10:46:05,926 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:46:05,926 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741977_1153 (size=68) 2025-01-08T10:46:05,926 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-1 {}] regionserver.HRegion(1681): Closing f2cb5b6ce155a147a1798fb975fd564d, disabling compactions & flushes 2025-01-08T10:46:05,926 INFO [RegionOpenAndInit-testtb-testExportFileSystemState-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d. 2025-01-08T10:46:05,926 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d. 2025-01-08T10:46:05,926 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d. after waiting 0 ms 2025-01-08T10:46:05,926 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d. 2025-01-08T10:46:05,926 INFO [RegionOpenAndInit-testtb-testExportFileSystemState-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d. 2025-01-08T10:46:05,926 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-1 {}] regionserver.HRegion(1635): Region close journal for f2cb5b6ce155a147a1798fb975fd564d: 2025-01-08T10:46:05,927 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:46:05,928 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-0 {}] regionserver.HRegion(1681): Closing 9f13bcd219cfe423da66381b35d7e9b8, disabling compactions & flushes 2025-01-08T10:46:05,928 INFO [RegionOpenAndInit-testtb-testExportFileSystemState-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8. 2025-01-08T10:46:05,928 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8. 2025-01-08T10:46:05,928 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8. after waiting 0 ms 2025-01-08T10:46:05,928 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8. 2025-01-08T10:46:05,928 INFO [RegionOpenAndInit-testtb-testExportFileSystemState-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8. 2025-01-08T10:46:05,928 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-0 {}] regionserver.HRegion(1635): Region close journal for 9f13bcd219cfe423da66381b35d7e9b8: 2025-01-08T10:46:05,929 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=63, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testExportFileSystemState execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T10:46:05,929 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d.","families":{"info":[{"qualifier":"regioninfo","vlen":67,"tag":[],"timestamp":"1736333165929"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736333165929"}]},"ts":"1736333165929"} 2025-01-08T10:46:05,929 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8.","families":{"info":[{"qualifier":"regioninfo","vlen":67,"tag":[],"timestamp":"1736333165929"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736333165929"}]},"ts":"1736333165929"} 2025-01-08T10:46:05,932 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T10:46:05,933 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=63, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testExportFileSystemState execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T10:46:05,933 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333165933"}]},"ts":"1736333165933"} 2025-01-08T10:46:05,934 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemState, state=ENABLING in hbase:meta 2025-01-08T10:46:05,938 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(202): Hosts are {67f42ee69b4e=0} racks are {/default-rack=0} 2025-01-08T10:46:05,940 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T10:46:05,940 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T10:46:05,940 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T10:46:05,940 INFO [PEWorker-5 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T10:46:05,940 INFO [PEWorker-5 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T10:46:05,940 INFO [PEWorker-5 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T10:46:05,940 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T10:46:05,940 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=64, ppid=63, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=9f13bcd219cfe423da66381b35d7e9b8, ASSIGN}, {pid=65, ppid=63, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=f2cb5b6ce155a147a1798fb975fd564d, ASSIGN}] 2025-01-08T10:46:05,941 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=65, ppid=63, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=f2cb5b6ce155a147a1798fb975fd564d, ASSIGN 2025-01-08T10:46:05,942 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=64, ppid=63, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=9f13bcd219cfe423da66381b35d7e9b8, ASSIGN 2025-01-08T10:46:05,942 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=65, ppid=63, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=f2cb5b6ce155a147a1798fb975fd564d, ASSIGN; state=OFFLINE, location=67f42ee69b4e,45003,1736333108039; forceNewPlan=false, retain=false 2025-01-08T10:46:05,943 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=64, ppid=63, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=9f13bcd219cfe423da66381b35d7e9b8, ASSIGN; state=OFFLINE, location=67f42ee69b4e,38323,1736333108259; forceNewPlan=false, retain=false 2025-01-08T10:46:05,996 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=63 2025-01-08T10:46:06,093 INFO [67f42ee69b4e:34551 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T10:46:06,093 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=64 updating hbase:meta row=9f13bcd219cfe423da66381b35d7e9b8, regionState=OPENING, regionLocation=67f42ee69b4e,38323,1736333108259 2025-01-08T10:46:06,093 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=65 updating hbase:meta row=f2cb5b6ce155a147a1798fb975fd564d, regionState=OPENING, regionLocation=67f42ee69b4e,45003,1736333108039 2025-01-08T10:46:06,096 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=66, ppid=64, state=RUNNABLE; OpenRegionProcedure 9f13bcd219cfe423da66381b35d7e9b8, server=67f42ee69b4e,38323,1736333108259}] 2025-01-08T10:46:06,099 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=67, ppid=65, state=RUNNABLE; OpenRegionProcedure f2cb5b6ce155a147a1798fb975fd564d, server=67f42ee69b4e,45003,1736333108039}] 2025-01-08T10:46:06,201 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=63 2025-01-08T10:46:06,252 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,38323,1736333108259 2025-01-08T10:46:06,252 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:46:06,256 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] handler.AssignRegionHandler(135): Open testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d. 2025-01-08T10:46:06,257 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(7285): Opening region: {ENCODED => f2cb5b6ce155a147a1798fb975fd564d, NAME => 'testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T10:46:06,257 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d. service=AccessControlService 2025-01-08T10:46:06,258 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:46:06,258 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportFileSystemState f2cb5b6ce155a147a1798fb975fd564d 2025-01-08T10:46:06,258 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:46:06,258 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(7327): checking encryption for f2cb5b6ce155a147a1798fb975fd564d 2025-01-08T10:46:06,258 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(7330): checking classloading for f2cb5b6ce155a147a1798fb975fd564d 2025-01-08T10:46:06,260 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] handler.AssignRegionHandler(135): Open testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8. 2025-01-08T10:46:06,260 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(7285): Opening region: {ENCODED => 9f13bcd219cfe423da66381b35d7e9b8, NAME => 'testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T10:46:06,260 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8. service=AccessControlService 2025-01-08T10:46:06,260 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:46:06,261 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportFileSystemState 9f13bcd219cfe423da66381b35d7e9b8 2025-01-08T10:46:06,261 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:46:06,261 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(7327): checking encryption for 9f13bcd219cfe423da66381b35d7e9b8 2025-01-08T10:46:06,261 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(7330): checking classloading for 9f13bcd219cfe423da66381b35d7e9b8 2025-01-08T10:46:06,262 INFO [StoreOpener-f2cb5b6ce155a147a1798fb975fd564d-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region f2cb5b6ce155a147a1798fb975fd564d 2025-01-08T10:46:06,264 INFO [StoreOpener-f2cb5b6ce155a147a1798fb975fd564d-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region f2cb5b6ce155a147a1798fb975fd564d columnFamilyName cf 2025-01-08T10:46:06,266 INFO [StoreOpener-9f13bcd219cfe423da66381b35d7e9b8-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 9f13bcd219cfe423da66381b35d7e9b8 2025-01-08T10:46:06,267 DEBUG [StoreOpener-f2cb5b6ce155a147a1798fb975fd564d-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:46:06,267 INFO [StoreOpener-f2cb5b6ce155a147a1798fb975fd564d-1 {}] regionserver.HStore(327): Store=f2cb5b6ce155a147a1798fb975fd564d/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:46:06,269 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/f2cb5b6ce155a147a1798fb975fd564d 2025-01-08T10:46:06,269 INFO [StoreOpener-9f13bcd219cfe423da66381b35d7e9b8-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 9f13bcd219cfe423da66381b35d7e9b8 columnFamilyName cf 2025-01-08T10:46:06,269 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/f2cb5b6ce155a147a1798fb975fd564d 2025-01-08T10:46:06,270 DEBUG [StoreOpener-9f13bcd219cfe423da66381b35d7e9b8-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:46:06,270 INFO [StoreOpener-9f13bcd219cfe423da66381b35d7e9b8-1 {}] regionserver.HStore(327): Store=9f13bcd219cfe423da66381b35d7e9b8/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:46:06,271 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/9f13bcd219cfe423da66381b35d7e9b8 2025-01-08T10:46:06,272 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/9f13bcd219cfe423da66381b35d7e9b8 2025-01-08T10:46:06,274 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(1085): writing seq id for f2cb5b6ce155a147a1798fb975fd564d 2025-01-08T10:46:06,275 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(1085): writing seq id for 9f13bcd219cfe423da66381b35d7e9b8 2025-01-08T10:46:06,277 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/f2cb5b6ce155a147a1798fb975fd564d/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:46:06,278 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(1102): Opened f2cb5b6ce155a147a1798fb975fd564d; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=69497276, jitterRate=0.03559011220932007}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T10:46:06,279 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(1001): Region open journal for f2cb5b6ce155a147a1798fb975fd564d: 2025-01-08T10:46:06,280 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d., pid=67, masterSystemTime=1736333166252 2025-01-08T10:46:06,282 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/9f13bcd219cfe423da66381b35d7e9b8/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:46:06,283 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d. 2025-01-08T10:46:06,283 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] handler.AssignRegionHandler(164): Opened testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d. 2025-01-08T10:46:06,286 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(1102): Opened 9f13bcd219cfe423da66381b35d7e9b8; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=70478500, jitterRate=0.05021148920059204}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T10:46:06,286 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(1001): Region open journal for 9f13bcd219cfe423da66381b35d7e9b8: 2025-01-08T10:46:06,287 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=65 updating hbase:meta row=f2cb5b6ce155a147a1798fb975fd564d, regionState=OPEN, openSeqNum=2, regionLocation=67f42ee69b4e,45003,1736333108039 2025-01-08T10:46:06,287 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8., pid=66, masterSystemTime=1736333166252 2025-01-08T10:46:06,291 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8. 2025-01-08T10:46:06,291 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] handler.AssignRegionHandler(164): Opened testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8. 2025-01-08T10:46:06,296 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=64 updating hbase:meta row=9f13bcd219cfe423da66381b35d7e9b8, regionState=OPEN, openSeqNum=2, regionLocation=67f42ee69b4e,38323,1736333108259 2025-01-08T10:46:06,299 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=67, resume processing ppid=65 2025-01-08T10:46:06,302 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=66, resume processing ppid=64 2025-01-08T10:46:06,302 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=66, ppid=64, state=SUCCESS; OpenRegionProcedure 9f13bcd219cfe423da66381b35d7e9b8, server=67f42ee69b4e,38323,1736333108259 in 204 msec 2025-01-08T10:46:06,302 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=65, ppid=63, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=f2cb5b6ce155a147a1798fb975fd564d, ASSIGN in 360 msec 2025-01-08T10:46:06,303 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=67, ppid=65, state=SUCCESS; OpenRegionProcedure f2cb5b6ce155a147a1798fb975fd564d, server=67f42ee69b4e,45003,1736333108039 in 195 msec 2025-01-08T10:46:06,305 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=64, resume processing ppid=63 2025-01-08T10:46:06,305 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=64, ppid=63, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=9f13bcd219cfe423da66381b35d7e9b8, ASSIGN in 362 msec 2025-01-08T10:46:06,306 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=63, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testExportFileSystemState execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T10:46:06,306 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333166306"}]},"ts":"1736333166306"} 2025-01-08T10:46:06,308 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemState, state=ENABLED in hbase:meta 2025-01-08T10:46:06,312 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=63, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportFileSystemState execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T10:46:06,312 DEBUG [PEWorker-2 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testExportFileSystemState jenkins: RWXCA 2025-01-08T10:46:06,315 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36283 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemState], kv [jenkins: RWXCA] 2025-01-08T10:46:06,317 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:46:06,317 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:46:06,317 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:46:06,318 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:46:06,320 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemState with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportFileSystemState \x00 \x01 \x02 \x03 \x04 2025-01-08T10:46:06,320 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemState with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportFileSystemState \x00 \x01 \x02 \x03 \x04 2025-01-08T10:46:06,321 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemState with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportFileSystemState \x00 \x01 \x02 \x03 \x04 2025-01-08T10:46:06,321 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemState with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportFileSystemState \x00 \x01 \x02 \x03 \x04 2025-01-08T10:46:06,325 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=63, state=SUCCESS; CreateTableProcedure table=testtb-testExportFileSystemState in 428 msec 2025-01-08T10:46:06,502 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=63 2025-01-08T10:46:06,503 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testExportFileSystemState, procId: 63 completed 2025-01-08T10:46:06,506 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportFileSystemState 2025-01-08T10:46:06,506 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8. 2025-01-08T10:46:06,507 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T10:46:06,516 DEBUG [hconnection-0x18c94e6b-shared-pool-16 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:46:06,519 INFO [RS-EventLoopGroup-5-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:46322, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:46:06,524 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.2 snapshot request for:{ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } 2025-01-08T10:46:06,524 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736333166524 (current time:1736333166524). 2025-01-08T10:46:06,524 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T10:46:06,524 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot emptySnaptb0-testExportFileSystemState VERSION not specified, setting to 2 2025-01-08T10:46:06,524 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T10:46:06,525 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x18f385e0 to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@7fd6b4f 2025-01-08T10:46:06,529 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@46205009, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:46:06,530 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:46:06,531 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:42850, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:46:06,533 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x18f385e0 to 127.0.0.1:52367 2025-01-08T10:46:06,533 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:46:06,534 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x6f7869df to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@6d8b8c1a 2025-01-08T10:46:06,537 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@296f29ba, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:46:06,538 DEBUG [hconnection-0x5c955762-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:46:06,539 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:42856, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:46:06,542 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x6f7869df to 127.0.0.1:52367 2025-01-08T10:46:06,542 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:46:06,542 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemState], kv [jenkins: RWXCA] 2025-01-08T10:46:06,543 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T10:46:06,544 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=68, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } 2025-01-08T10:46:06,544 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 }, snapshot procedure id = 68 2025-01-08T10:46:06,545 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T10:46:06,546 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=68 2025-01-08T10:46:06,546 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T10:46:06,549 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T10:46:06,561 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741979_1155 (size=170) 2025-01-08T10:46:06,561 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741979_1155 (size=170) 2025-01-08T10:46:06,562 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741979_1155 (size=170) 2025-01-08T10:46:06,563 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T10:46:06,564 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=69, ppid=68, state=RUNNABLE; SnapshotRegionProcedure 9f13bcd219cfe423da66381b35d7e9b8}, {pid=70, ppid=68, state=RUNNABLE; SnapshotRegionProcedure f2cb5b6ce155a147a1798fb975fd564d}] 2025-01-08T10:46:06,565 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=69, ppid=68, state=RUNNABLE; SnapshotRegionProcedure 9f13bcd219cfe423da66381b35d7e9b8 2025-01-08T10:46:06,565 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=70, ppid=68, state=RUNNABLE; SnapshotRegionProcedure f2cb5b6ce155a147a1798fb975fd564d 2025-01-08T10:46:06,647 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=68 2025-01-08T10:46:06,716 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:46:06,716 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,38323,1736333108259 2025-01-08T10:46:06,716 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=45003 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=70 2025-01-08T10:46:06,716 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=38323 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=69 2025-01-08T10:46:06,716 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=70}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d. 2025-01-08T10:46:06,717 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=69}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8. 2025-01-08T10:46:06,717 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=70}] regionserver.HRegion(2538): Flush status journal for f2cb5b6ce155a147a1798fb975fd564d: 2025-01-08T10:46:06,717 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=70}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d. for emptySnaptb0-testExportFileSystemState completed. 2025-01-08T10:46:06,717 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=69}] regionserver.HRegion(2538): Flush status journal for 9f13bcd219cfe423da66381b35d7e9b8: 2025-01-08T10:46:06,717 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=69}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8. for emptySnaptb0-testExportFileSystemState completed. 2025-01-08T10:46:06,717 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=70}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d.' region-info for snapshot=emptySnaptb0-testExportFileSystemState 2025-01-08T10:46:06,717 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=69}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8.' region-info for snapshot=emptySnaptb0-testExportFileSystemState 2025-01-08T10:46:06,717 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=70}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:46:06,717 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=69}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:46:06,717 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=70}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T10:46:06,717 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=69}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T10:46:06,725 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741980_1156 (size=71) 2025-01-08T10:46:06,725 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741980_1156 (size=71) 2025-01-08T10:46:06,725 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741980_1156 (size=71) 2025-01-08T10:46:06,726 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=70}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d. 2025-01-08T10:46:06,726 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=70}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=70 2025-01-08T10:46:06,726 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=70 2025-01-08T10:46:06,726 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportFileSystemState on region f2cb5b6ce155a147a1798fb975fd564d 2025-01-08T10:46:06,727 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=70, ppid=68, state=RUNNABLE; SnapshotRegionProcedure f2cb5b6ce155a147a1798fb975fd564d 2025-01-08T10:46:06,729 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=70, ppid=68, state=SUCCESS; SnapshotRegionProcedure f2cb5b6ce155a147a1798fb975fd564d in 163 msec 2025-01-08T10:46:06,734 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741981_1157 (size=71) 2025-01-08T10:46:06,734 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741981_1157 (size=71) 2025-01-08T10:46:06,735 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741981_1157 (size=71) 2025-01-08T10:46:06,735 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=69}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8. 2025-01-08T10:46:06,735 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=69}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=69 2025-01-08T10:46:06,736 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=69 2025-01-08T10:46:06,736 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportFileSystemState on region 9f13bcd219cfe423da66381b35d7e9b8 2025-01-08T10:46:06,736 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=69, ppid=68, state=RUNNABLE; SnapshotRegionProcedure 9f13bcd219cfe423da66381b35d7e9b8 2025-01-08T10:46:06,738 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=69, resume processing ppid=68 2025-01-08T10:46:06,738 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=69, ppid=68, state=SUCCESS; SnapshotRegionProcedure 9f13bcd219cfe423da66381b35d7e9b8 in 173 msec 2025-01-08T10:46:06,738 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T10:46:06,739 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T10:46:06,740 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T10:46:06,740 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T10:46:06,740 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:46:06,741 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(328): No files under family: cf 2025-01-08T10:46:06,750 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741982_1158 (size=63) 2025-01-08T10:46:06,750 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741982_1158 (size=63) 2025-01-08T10:46:06,751 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741982_1158 (size=63) 2025-01-08T10:46:06,752 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T10:46:06,752 DEBUG [PEWorker-4 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for emptySnaptb0-testExportFileSystemState 2025-01-08T10:46:06,753 DEBUG [PEWorker-4 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/emptySnaptb0-testExportFileSystemState 2025-01-08T10:46:06,762 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741983_1159 (size=653) 2025-01-08T10:46:06,762 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741983_1159 (size=653) 2025-01-08T10:46:06,763 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741983_1159 (size=653) 2025-01-08T10:46:06,770 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T10:46:06,775 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T10:46:06,776 DEBUG [PEWorker-4 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/emptySnaptb0-testExportFileSystemState to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/emptySnaptb0-testExportFileSystemState 2025-01-08T10:46:06,777 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T10:46:06,777 DEBUG [PEWorker-4 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 }, snapshot procedure id = 68 2025-01-08T10:46:06,778 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=68, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } in 234 msec 2025-01-08T10:46:06,848 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=68 2025-01-08T10:46:06,848 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportFileSystemState, procId: 68 completed 2025-01-08T10:46:06,851 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38323 {}] regionserver.HRegion(8254): writing data to region testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T10:46:06,853 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=45003 {}] regionserver.HRegion(8254): writing data to region testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T10:46:06,856 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportFileSystemState 2025-01-08T10:46:06,856 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8. 2025-01-08T10:46:06,856 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T10:46:06,869 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.2 snapshot request for:{ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } 2025-01-08T10:46:06,869 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736333166869 (current time:1736333166869). 2025-01-08T10:46:06,869 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T10:46:06,869 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testExportFileSystemState VERSION not specified, setting to 2 2025-01-08T10:46:06,869 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T10:46:06,870 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x11278190 to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@67ffea91 2025-01-08T10:46:06,873 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@68a2afdb, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:46:06,875 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:46:06,876 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:42866, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:46:06,876 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x11278190 to 127.0.0.1:52367 2025-01-08T10:46:06,876 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:46:06,877 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x2e0d40dc to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@36be9cee 2025-01-08T10:46:06,880 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4b9f2195, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:46:06,882 DEBUG [hconnection-0x139de8d0-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:46:06,883 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:42872, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:46:06,885 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x2e0d40dc to 127.0.0.1:52367 2025-01-08T10:46:06,885 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:46:06,885 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemState], kv [jenkins: RWXCA] 2025-01-08T10:46:06,886 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T10:46:06,887 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=71, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } 2025-01-08T10:46:06,887 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 }, snapshot procedure id = 71 2025-01-08T10:46:06,888 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T10:46:06,890 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T10:46:06,890 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=71 2025-01-08T10:46:06,892 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T10:46:06,907 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741984_1160 (size=165) 2025-01-08T10:46:06,908 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741984_1160 (size=165) 2025-01-08T10:46:06,908 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741984_1160 (size=165) 2025-01-08T10:46:06,915 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T10:46:06,915 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=72, ppid=71, state=RUNNABLE; SnapshotRegionProcedure 9f13bcd219cfe423da66381b35d7e9b8}, {pid=73, ppid=71, state=RUNNABLE; SnapshotRegionProcedure f2cb5b6ce155a147a1798fb975fd564d}] 2025-01-08T10:46:06,916 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=72, ppid=71, state=RUNNABLE; SnapshotRegionProcedure 9f13bcd219cfe423da66381b35d7e9b8 2025-01-08T10:46:06,916 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=73, ppid=71, state=RUNNABLE; SnapshotRegionProcedure f2cb5b6ce155a147a1798fb975fd564d 2025-01-08T10:46:06,991 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=71 2025-01-08T10:46:07,066 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,38323,1736333108259 2025-01-08T10:46:07,067 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=38323 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=72 2025-01-08T10:46:07,067 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:46:07,067 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8. 2025-01-08T10:46:07,067 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.HRegion(2837): Flushing 9f13bcd219cfe423da66381b35d7e9b8 1/1 column families, dataSize=333 B heapSize=976 B 2025-01-08T10:46:07,067 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=45003 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=73 2025-01-08T10:46:07,068 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d. 2025-01-08T10:46:07,068 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.HRegion(2837): Flushing f2cb5b6ce155a147a1798fb975fd564d 1/1 column families, dataSize=2.93 KB heapSize=6.58 KB 2025-01-08T10:46:07,087 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e202501087e69e94c8e224ddcaab3c0b8e50fb119_9f13bcd219cfe423da66381b35d7e9b8 is 71, key is 09cb6af555951dcbed9f7a47688642e9/cf:q/1736333166851/Put/seqid=0 2025-01-08T10:46:07,093 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741985_1161 (size=5242) 2025-01-08T10:46:07,094 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741985_1161 (size=5242) 2025-01-08T10:46:07,094 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741985_1161 (size=5242) 2025-01-08T10:46:07,094 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b202501085f987b4e97ca4ffe8e71a63b1de64287_f2cb5b6ce155a147a1798fb975fd564d is 71, key is 11b56d6af0672cb5d00f730c1006019a/cf:q/1736333166852/Put/seqid=0 2025-01-08T10:46:07,095 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:46:07,101 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741986_1162 (size=8031) 2025-01-08T10:46:07,102 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741986_1162 (size=8031) 2025-01-08T10:46:07,103 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741986_1162 (size=8031) 2025-01-08T10:46:07,103 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:46:07,111 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b202501085f987b4e97ca4ffe8e71a63b1de64287_f2cb5b6ce155a147a1798fb975fd564d to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/c4ca4238a0b923820dcc509a6f75849b202501085f987b4e97ca4ffe8e71a63b1de64287_f2cb5b6ce155a147a1798fb975fd564d 2025-01-08T10:46:07,112 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/f2cb5b6ce155a147a1798fb975fd564d/.tmp/cf/b50639db54bf4cffa375e391dca52f5c, store: [table=testtb-testExportFileSystemState family=cf region=f2cb5b6ce155a147a1798fb975fd564d] 2025-01-08T10:46:07,113 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/f2cb5b6ce155a147a1798fb975fd564d/.tmp/cf/b50639db54bf4cffa375e391dca52f5c is 209, key is 1e2a7474ba3750c70956b6fe867caaf52/cf:q/1736333166852/Put/seqid=0 2025-01-08T10:46:07,118 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e202501087e69e94c8e224ddcaab3c0b8e50fb119_9f13bcd219cfe423da66381b35d7e9b8 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/d41d8cd98f00b204e9800998ecf8427e202501087e69e94c8e224ddcaab3c0b8e50fb119_9f13bcd219cfe423da66381b35d7e9b8 2025-01-08T10:46:07,119 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741987_1163 (size=14587) 2025-01-08T10:46:07,120 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/9f13bcd219cfe423da66381b35d7e9b8/.tmp/cf/e9d58b0004864757b4fde3925177d46d, store: [table=testtb-testExportFileSystemState family=cf region=9f13bcd219cfe423da66381b35d7e9b8] 2025-01-08T10:46:07,120 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741987_1163 (size=14587) 2025-01-08T10:46:07,121 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/9f13bcd219cfe423da66381b35d7e9b8/.tmp/cf/e9d58b0004864757b4fde3925177d46d is 209, key is 0cb71474a2691412d0fed723b81df666d/cf:q/1736333166851/Put/seqid=0 2025-01-08T10:46:07,121 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741987_1163 (size=14587) 2025-01-08T10:46:07,122 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=2.9 K, hasBloomFilter=true, into tmp file hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/f2cb5b6ce155a147a1798fb975fd564d/.tmp/cf/b50639db54bf4cffa375e391dca52f5c 2025-01-08T10:46:07,128 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741988_1164 (size=6328) 2025-01-08T10:46:07,128 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741988_1164 (size=6328) 2025-01-08T10:46:07,129 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/f2cb5b6ce155a147a1798fb975fd564d/.tmp/cf/b50639db54bf4cffa375e391dca52f5c as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/f2cb5b6ce155a147a1798fb975fd564d/cf/b50639db54bf4cffa375e391dca52f5c 2025-01-08T10:46:07,129 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741988_1164 (size=6328) 2025-01-08T10:46:07,129 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=333, hasBloomFilter=true, into tmp file hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/9f13bcd219cfe423da66381b35d7e9b8/.tmp/cf/e9d58b0004864757b4fde3925177d46d 2025-01-08T10:46:07,134 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/f2cb5b6ce155a147a1798fb975fd564d/cf/b50639db54bf4cffa375e391dca52f5c, entries=45, sequenceid=6, filesize=14.2 K 2025-01-08T10:46:07,135 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/9f13bcd219cfe423da66381b35d7e9b8/.tmp/cf/e9d58b0004864757b4fde3925177d46d as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/9f13bcd219cfe423da66381b35d7e9b8/cf/e9d58b0004864757b4fde3925177d46d 2025-01-08T10:46:07,136 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.HRegion(3040): Finished flush of dataSize ~2.93 KB/3003, heapSize ~6.56 KB/6720, currentSize=0 B/0 for f2cb5b6ce155a147a1798fb975fd564d in 67ms, sequenceid=6, compaction requested=false 2025-01-08T10:46:07,136 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportFileSystemState' 2025-01-08T10:46:07,136 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.HRegion(2538): Flush status journal for f2cb5b6ce155a147a1798fb975fd564d: 2025-01-08T10:46:07,136 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d. for snaptb0-testExportFileSystemState completed. 2025-01-08T10:46:07,137 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d.' region-info for snapshot=snaptb0-testExportFileSystemState 2025-01-08T10:46:07,137 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:46:07,137 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/f2cb5b6ce155a147a1798fb975fd564d/cf/b50639db54bf4cffa375e391dca52f5c] hfiles 2025-01-08T10:46:07,137 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/f2cb5b6ce155a147a1798fb975fd564d/cf/b50639db54bf4cffa375e391dca52f5c for snapshot=snaptb0-testExportFileSystemState 2025-01-08T10:46:07,141 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/9f13bcd219cfe423da66381b35d7e9b8/cf/e9d58b0004864757b4fde3925177d46d, entries=5, sequenceid=6, filesize=6.2 K 2025-01-08T10:46:07,142 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.HRegion(3040): Finished flush of dataSize ~333 B/333, heapSize ~960 B/960, currentSize=0 B/0 for 9f13bcd219cfe423da66381b35d7e9b8 in 75ms, sequenceid=6, compaction requested=false 2025-01-08T10:46:07,142 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.HRegion(2538): Flush status journal for 9f13bcd219cfe423da66381b35d7e9b8: 2025-01-08T10:46:07,142 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8. for snaptb0-testExportFileSystemState completed. 2025-01-08T10:46:07,142 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8.' region-info for snapshot=snaptb0-testExportFileSystemState 2025-01-08T10:46:07,142 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:46:07,142 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/9f13bcd219cfe423da66381b35d7e9b8/cf/e9d58b0004864757b4fde3925177d46d] hfiles 2025-01-08T10:46:07,142 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/9f13bcd219cfe423da66381b35d7e9b8/cf/e9d58b0004864757b4fde3925177d46d for snapshot=snaptb0-testExportFileSystemState 2025-01-08T10:46:07,144 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741989_1165 (size=110) 2025-01-08T10:46:07,145 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741989_1165 (size=110) 2025-01-08T10:46:07,145 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d. 2025-01-08T10:46:07,145 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=73 2025-01-08T10:46:07,145 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741989_1165 (size=110) 2025-01-08T10:46:07,146 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=73 2025-01-08T10:46:07,146 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportFileSystemState on region f2cb5b6ce155a147a1798fb975fd564d 2025-01-08T10:46:07,146 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=73, ppid=71, state=RUNNABLE; SnapshotRegionProcedure f2cb5b6ce155a147a1798fb975fd564d 2025-01-08T10:46:07,149 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=73, ppid=71, state=SUCCESS; SnapshotRegionProcedure f2cb5b6ce155a147a1798fb975fd564d in 232 msec 2025-01-08T10:46:07,154 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741990_1166 (size=110) 2025-01-08T10:46:07,154 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741990_1166 (size=110) 2025-01-08T10:46:07,154 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741990_1166 (size=110) 2025-01-08T10:46:07,155 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8. 2025-01-08T10:46:07,155 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=72 2025-01-08T10:46:07,155 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=72 2025-01-08T10:46:07,156 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportFileSystemState on region 9f13bcd219cfe423da66381b35d7e9b8 2025-01-08T10:46:07,156 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=72, ppid=71, state=RUNNABLE; SnapshotRegionProcedure 9f13bcd219cfe423da66381b35d7e9b8 2025-01-08T10:46:07,158 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=72, resume processing ppid=71 2025-01-08T10:46:07,158 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=72, ppid=71, state=SUCCESS; SnapshotRegionProcedure 9f13bcd219cfe423da66381b35d7e9b8 in 241 msec 2025-01-08T10:46:07,158 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T10:46:07,159 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T10:46:07,160 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T10:46:07,160 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T10:46:07,160 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:46:07,161 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/c4ca4238a0b923820dcc509a6f75849b202501085f987b4e97ca4ffe8e71a63b1de64287_f2cb5b6ce155a147a1798fb975fd564d, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/d41d8cd98f00b204e9800998ecf8427e202501087e69e94c8e224ddcaab3c0b8e50fb119_9f13bcd219cfe423da66381b35d7e9b8] hfiles 2025-01-08T10:46:07,161 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/c4ca4238a0b923820dcc509a6f75849b202501085f987b4e97ca4ffe8e71a63b1de64287_f2cb5b6ce155a147a1798fb975fd564d 2025-01-08T10:46:07,161 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/d41d8cd98f00b204e9800998ecf8427e202501087e69e94c8e224ddcaab3c0b8e50fb119_9f13bcd219cfe423da66381b35d7e9b8 2025-01-08T10:46:07,168 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741991_1167 (size=294) 2025-01-08T10:46:07,168 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741991_1167 (size=294) 2025-01-08T10:46:07,168 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741991_1167 (size=294) 2025-01-08T10:46:07,169 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T10:46:07,169 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testExportFileSystemState 2025-01-08T10:46:07,170 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemState 2025-01-08T10:46:07,179 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741992_1168 (size=963) 2025-01-08T10:46:07,179 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741992_1168 (size=963) 2025-01-08T10:46:07,180 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741992_1168 (size=963) 2025-01-08T10:46:07,183 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T10:46:07,189 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T10:46:07,189 DEBUG [PEWorker-1 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemState to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportFileSystemState 2025-01-08T10:46:07,190 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T10:46:07,191 DEBUG [PEWorker-1 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 }, snapshot procedure id = 71 2025-01-08T10:46:07,192 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=71, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } in 305 msec 2025-01-08T10:46:07,192 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=71 2025-01-08T10:46:07,192 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportFileSystemState, procId: 71 completed 2025-01-08T10:46:07,193 INFO [Time-limited test {}] snapshot.TestExportSnapshot(468): HDFS export destination path: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333167192 2025-01-08T10:46:07,193 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=hdfs://localhost:37495, tgtDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333167192, rawTgtDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333167192, srcFsUri=hdfs://localhost:37495, srcDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:46:07,222 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:37495, inputRoot=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:46:07,222 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-832679797_22, ugi=jenkins (auth:SIMPLE)]], outputRoot=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333167192, skipTmp=false, initialOutputSnapshotDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333167192/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemState 2025-01-08T10:46:07,224 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T10:46:07,228 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportFileSystemState to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333167192/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemState 2025-01-08T10:46:07,244 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741993_1169 (size=165) 2025-01-08T10:46:07,244 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741993_1169 (size=165) 2025-01-08T10:46:07,244 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741993_1169 (size=165) 2025-01-08T10:46:07,252 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741994_1170 (size=963) 2025-01-08T10:46:07,252 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741994_1170 (size=963) 2025-01-08T10:46:07,253 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741994_1170 (size=963) 2025-01-08T10:46:07,429 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-9429268103549186973.jar 2025-01-08T10:46:07,429 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:07,430 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:07,430 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:07,595 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemState 2025-01-08T10:46:07,596 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemState Metrics about Tables on a single HBase RegionServer 2025-01-08T10:46:07,596 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testExportWithResetTtl 2025-01-08T10:46:07,597 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportWithResetTtl 2025-01-08T10:46:08,451 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-17984237517410454201.jar 2025-01-08T10:46:08,452 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:08,452 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:08,520 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-2674273912991722708.jar 2025-01-08T10:46:08,521 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:08,521 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:08,521 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:08,521 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:08,522 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:08,522 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:08,522 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T10:46:08,522 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T10:46:08,523 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T10:46:08,523 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T10:46:08,523 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T10:46:08,523 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T10:46:08,524 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T10:46:08,524 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T10:46:08,524 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T10:46:08,524 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T10:46:08,524 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T10:46:08,525 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T10:46:08,525 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:46:08,525 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:46:08,526 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:46:08,526 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:46:08,526 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:46:08,527 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:46:08,527 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:46:08,583 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741995_1171 (size=127628) 2025-01-08T10:46:08,583 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741995_1171 (size=127628) 2025-01-08T10:46:08,583 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741995_1171 (size=127628) 2025-01-08T10:46:08,603 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741996_1172 (size=2172137) 2025-01-08T10:46:08,603 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741996_1172 (size=2172137) 2025-01-08T10:46:08,603 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741996_1172 (size=2172137) 2025-01-08T10:46:09,011 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741997_1173 (size=213228) 2025-01-08T10:46:09,012 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741997_1173 (size=213228) 2025-01-08T10:46:09,012 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741997_1173 (size=213228) 2025-01-08T10:46:09,020 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741998_1174 (size=451756) 2025-01-08T10:46:09,020 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741998_1174 (size=451756) 2025-01-08T10:46:09,021 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741998_1174 (size=451756) 2025-01-08T10:46:09,033 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741999_1175 (size=1877034) 2025-01-08T10:46:09,033 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741999_1175 (size=1877034) 2025-01-08T10:46:09,033 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741999_1175 (size=1877034) 2025-01-08T10:46:09,060 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742000_1176 (size=6350714) 2025-01-08T10:46:09,060 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742000_1176 (size=6350714) 2025-01-08T10:46:09,061 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742000_1176 (size=6350714) 2025-01-08T10:46:09,070 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742001_1177 (size=533455) 2025-01-08T10:46:09,070 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742001_1177 (size=533455) 2025-01-08T10:46:09,071 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742001_1177 (size=533455) 2025-01-08T10:46:09,098 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742002_1178 (size=7280644) 2025-01-08T10:46:09,098 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742002_1178 (size=7280644) 2025-01-08T10:46:09,099 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742002_1178 (size=7280644) 2025-01-08T10:46:09,135 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742003_1179 (size=4188619) 2025-01-08T10:46:09,135 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742003_1179 (size=4188619) 2025-01-08T10:46:09,136 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742003_1179 (size=4188619) 2025-01-08T10:46:09,147 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742004_1180 (size=20406) 2025-01-08T10:46:09,147 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742004_1180 (size=20406) 2025-01-08T10:46:09,147 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742004_1180 (size=20406) 2025-01-08T10:46:09,156 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742005_1181 (size=75495) 2025-01-08T10:46:09,156 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742005_1181 (size=75495) 2025-01-08T10:46:09,157 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742005_1181 (size=75495) 2025-01-08T10:46:09,166 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742006_1182 (size=45609) 2025-01-08T10:46:09,166 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742006_1182 (size=45609) 2025-01-08T10:46:09,167 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742006_1182 (size=45609) 2025-01-08T10:46:09,174 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742007_1183 (size=110084) 2025-01-08T10:46:09,174 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742007_1183 (size=110084) 2025-01-08T10:46:09,174 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742007_1183 (size=110084) 2025-01-08T10:46:09,184 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742008_1184 (size=912094) 2025-01-08T10:46:09,184 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742008_1184 (size=912094) 2025-01-08T10:46:09,185 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742008_1184 (size=912094) 2025-01-08T10:46:09,197 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742009_1185 (size=1323991) 2025-01-08T10:46:09,197 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742009_1185 (size=1323991) 2025-01-08T10:46:09,197 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742009_1185 (size=1323991) 2025-01-08T10:46:09,205 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742010_1186 (size=23076) 2025-01-08T10:46:09,205 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742010_1186 (size=23076) 2025-01-08T10:46:09,205 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742010_1186 (size=23076) 2025-01-08T10:46:09,214 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742011_1187 (size=126803) 2025-01-08T10:46:09,214 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742011_1187 (size=126803) 2025-01-08T10:46:09,215 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742011_1187 (size=126803) 2025-01-08T10:46:09,224 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742012_1188 (size=322274) 2025-01-08T10:46:09,224 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742012_1188 (size=322274) 2025-01-08T10:46:09,224 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742012_1188 (size=322274) 2025-01-08T10:46:09,237 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742013_1189 (size=1832290) 2025-01-08T10:46:09,237 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742013_1189 (size=1832290) 2025-01-08T10:46:09,237 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742013_1189 (size=1832290) 2025-01-08T10:46:09,245 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742014_1190 (size=30081) 2025-01-08T10:46:09,245 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742014_1190 (size=30081) 2025-01-08T10:46:09,246 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742014_1190 (size=30081) 2025-01-08T10:46:09,253 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742015_1191 (size=53616) 2025-01-08T10:46:09,254 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742015_1191 (size=53616) 2025-01-08T10:46:09,254 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742015_1191 (size=53616) 2025-01-08T10:46:09,262 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742016_1192 (size=29229) 2025-01-08T10:46:09,262 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742016_1192 (size=29229) 2025-01-08T10:46:09,263 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742016_1192 (size=29229) 2025-01-08T10:46:09,272 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742017_1193 (size=169089) 2025-01-08T10:46:09,272 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742017_1193 (size=169089) 2025-01-08T10:46:09,272 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742017_1193 (size=169089) 2025-01-08T10:46:09,294 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742018_1194 (size=5175431) 2025-01-08T10:46:09,295 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742018_1194 (size=5175431) 2025-01-08T10:46:09,295 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742018_1194 (size=5175431) 2025-01-08T10:46:09,303 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742019_1195 (size=136454) 2025-01-08T10:46:09,304 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742019_1195 (size=136454) 2025-01-08T10:46:09,304 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742019_1195 (size=136454) 2025-01-08T10:46:09,318 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742020_1196 (size=3317408) 2025-01-08T10:46:09,319 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742020_1196 (size=3317408) 2025-01-08T10:46:09,319 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742020_1196 (size=3317408) 2025-01-08T10:46:09,328 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742021_1197 (size=503880) 2025-01-08T10:46:09,329 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742021_1197 (size=503880) 2025-01-08T10:46:09,329 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742021_1197 (size=503880) 2025-01-08T10:46:09,353 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742022_1198 (size=4695811) 2025-01-08T10:46:09,353 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742022_1198 (size=4695811) 2025-01-08T10:46:09,353 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742022_1198 (size=4695811) 2025-01-08T10:46:09,355 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T10:46:09,358 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'snaptb0-testExportFileSystemState' hfile list 2025-01-08T10:46:09,360 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(751): export split=0 size=33.4 K 2025-01-08T10:46:09,367 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742023_1199 (size=726) 2025-01-08T10:46:09,367 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742023_1199 (size=726) 2025-01-08T10:46:09,371 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742023_1199 (size=726) 2025-01-08T10:46:09,377 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742024_1200 (size=15) 2025-01-08T10:46:09,378 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742024_1200 (size=15) 2025-01-08T10:46:09,378 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742024_1200 (size=15) 2025-01-08T10:46:09,397 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742025_1201 (size=304935) 2025-01-08T10:46:09,397 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742025_1201 (size=304935) 2025-01-08T10:46:09,398 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742025_1201 (size=304935) 2025-01-08T10:46:09,813 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T10:46:09,813 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T10:46:09,816 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0002_000001 (auth:SIMPLE) from 127.0.0.1:39270 2025-01-08T10:46:09,833 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_1/usercache/jenkins/appcache/application_1736333116683_0002/container_1736333116683_0002_01_000001/launch_container.sh] 2025-01-08T10:46:09,833 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_1/usercache/jenkins/appcache/application_1736333116683_0002/container_1736333116683_0002_01_000001/container_tokens] 2025-01-08T10:46:09,834 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_1/usercache/jenkins/appcache/application_1736333116683_0002/container_1736333116683_0002_01_000001/sysfs] 2025-01-08T10:46:10,507 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0003_000001 (auth:SIMPLE) from 127.0.0.1:49464 2025-01-08T10:46:10,658 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T10:46:11,333 DEBUG [master/67f42ee69b4e:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 0c62df083bd10295a5e532128fcb0af8 changed from -1.0 to 0.0, refreshing cache 2025-01-08T10:46:11,333 DEBUG [master/67f42ee69b4e:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 873898d3d33a2f6df797594f02de3737 changed from -1.0 to 0.0, refreshing cache 2025-01-08T10:46:11,333 DEBUG [master/67f42ee69b4e:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 9f13bcd219cfe423da66381b35d7e9b8 changed from -1.0 to 0.0, refreshing cache 2025-01-08T10:46:11,334 DEBUG [master/67f42ee69b4e:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region f2cb5b6ce155a147a1798fb975fd564d changed from -1.0 to 0.0, refreshing cache 2025-01-08T10:46:15,848 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0003_000001 (auth:SIMPLE) from 127.0.0.1:43752 2025-01-08T10:46:16,122 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742026_1202 (size=350609) 2025-01-08T10:46:16,123 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742026_1202 (size=350609) 2025-01-08T10:46:16,123 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742026_1202 (size=350609) 2025-01-08T10:46:18,118 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0003_000001 (auth:SIMPLE) from 127.0.0.1:38956 2025-01-08T10:46:21,437 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742027_1203 (size=14587) 2025-01-08T10:46:21,437 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742027_1203 (size=14587) 2025-01-08T10:46:21,437 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742027_1203 (size=14587) 2025-01-08T10:46:21,500 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742028_1204 (size=8031) 2025-01-08T10:46:21,501 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742028_1204 (size=8031) 2025-01-08T10:46:21,501 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742028_1204 (size=8031) 2025-01-08T10:46:21,543 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742029_1205 (size=6328) 2025-01-08T10:46:21,543 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742029_1205 (size=6328) 2025-01-08T10:46:21,543 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742029_1205 (size=6328) 2025-01-08T10:46:21,590 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742030_1206 (size=5242) 2025-01-08T10:46:21,590 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742030_1206 (size=5242) 2025-01-08T10:46:21,590 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742030_1206 (size=5242) 2025-01-08T10:46:21,659 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742031_1207 (size=17462) 2025-01-08T10:46:21,660 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742031_1207 (size=17462) 2025-01-08T10:46:21,660 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742031_1207 (size=17462) 2025-01-08T10:46:21,690 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742032_1208 (size=465) 2025-01-08T10:46:21,691 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742032_1208 (size=465) 2025-01-08T10:46:21,691 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742032_1208 (size=465) 2025-01-08T10:46:21,727 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742033_1209 (size=17462) 2025-01-08T10:46:21,728 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742033_1209 (size=17462) 2025-01-08T10:46:21,728 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742033_1209 (size=17462) 2025-01-08T10:46:21,756 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742034_1210 (size=350609) 2025-01-08T10:46:21,756 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742034_1210 (size=350609) 2025-01-08T10:46:21,757 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742034_1210 (size=350609) 2025-01-08T10:46:21,776 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0003_000001 (auth:SIMPLE) from 127.0.0.1:38970 2025-01-08T10:46:21,787 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_2/usercache/jenkins/appcache/application_1736333116683_0003/container_1736333116683_0003_01_000002/launch_container.sh] 2025-01-08T10:46:21,787 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_2/usercache/jenkins/appcache/application_1736333116683_0003/container_1736333116683_0003_01_000002/container_tokens] 2025-01-08T10:46:21,787 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_2/usercache/jenkins/appcache/application_1736333116683_0003/container_1736333116683_0003_01_000002/sysfs] 2025-01-08T10:46:23,600 INFO [Time-limited test {}] snapshot.ExportSnapshot(1207): Finalize the Snapshot Export 2025-01-08T10:46:23,603 INFO [Time-limited test {}] snapshot.ExportSnapshot(1218): Verify the exported snapshot's expiration status and integrity. 2025-01-08T10:46:23,616 INFO [Time-limited test {}] snapshot.ExportSnapshot(1224): Export Completed: snaptb0-testExportFileSystemState 2025-01-08T10:46:23,616 INFO [Time-limited test {}] snapshot.TestExportSnapshot(362): Exported snapshot 2025-01-08T10:46:23,616 INFO [Time-limited test {}] snapshot.TestExportSnapshot(373): Verified filesystem state 2025-01-08T10:46:23,617 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-832679797_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportFileSystemState at hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportFileSystemState 2025-01-08T10:46:23,617 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportFileSystemState/.snapshotinfo 2025-01-08T10:46:23,617 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportFileSystemState/data.manifest 2025-01-08T10:46:23,617 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-832679797_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333167192/.hbase-snapshot/snaptb0-testExportFileSystemState at hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333167192/.hbase-snapshot/snaptb0-testExportFileSystemState 2025-01-08T10:46:23,618 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333167192/.hbase-snapshot/snaptb0-testExportFileSystemState/.snapshotinfo 2025-01-08T10:46:23,618 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333167192/.hbase-snapshot/snaptb0-testExportFileSystemState/data.manifest 2025-01-08T10:46:23,630 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testExportFileSystemState 2025-01-08T10:46:23,631 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.2 disable testtb-testExportFileSystemState 2025-01-08T10:46:23,634 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=74, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testExportFileSystemState 2025-01-08T10:46:23,639 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=74 2025-01-08T10:46:23,639 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333183639"}]},"ts":"1736333183639"} 2025-01-08T10:46:23,641 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemState, state=DISABLING in hbase:meta 2025-01-08T10:46:23,667 INFO [PEWorker-3 {}] procedure.DisableTableProcedure(284): Set testtb-testExportFileSystemState to state=DISABLING 2025-01-08T10:46:23,668 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=75, ppid=74, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testExportFileSystemState}] 2025-01-08T10:46:23,670 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=76, ppid=75, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=9f13bcd219cfe423da66381b35d7e9b8, UNASSIGN}, {pid=77, ppid=75, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=f2cb5b6ce155a147a1798fb975fd564d, UNASSIGN}] 2025-01-08T10:46:23,671 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=77, ppid=75, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=f2cb5b6ce155a147a1798fb975fd564d, UNASSIGN 2025-01-08T10:46:23,671 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=76, ppid=75, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=9f13bcd219cfe423da66381b35d7e9b8, UNASSIGN 2025-01-08T10:46:23,672 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=77 updating hbase:meta row=f2cb5b6ce155a147a1798fb975fd564d, regionState=CLOSING, regionLocation=67f42ee69b4e,45003,1736333108039 2025-01-08T10:46:23,672 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=76 updating hbase:meta row=9f13bcd219cfe423da66381b35d7e9b8, regionState=CLOSING, regionLocation=67f42ee69b4e,38323,1736333108259 2025-01-08T10:46:23,673 DEBUG [PEWorker-2 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T10:46:23,673 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=78, ppid=77, state=RUNNABLE; CloseRegionProcedure f2cb5b6ce155a147a1798fb975fd564d, server=67f42ee69b4e,45003,1736333108039}] 2025-01-08T10:46:23,674 DEBUG [PEWorker-4 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T10:46:23,674 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=79, ppid=76, state=RUNNABLE; CloseRegionProcedure 9f13bcd219cfe423da66381b35d7e9b8, server=67f42ee69b4e,38323,1736333108259}] 2025-01-08T10:46:23,740 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=74 2025-01-08T10:46:23,825 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:46:23,826 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] handler.UnassignRegionHandler(124): Close f2cb5b6ce155a147a1798fb975fd564d 2025-01-08T10:46:23,826 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T10:46:23,826 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,38323,1736333108259 2025-01-08T10:46:23,826 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] regionserver.HRegion(1681): Closing f2cb5b6ce155a147a1798fb975fd564d, disabling compactions & flushes 2025-01-08T10:46:23,826 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d. 2025-01-08T10:46:23,826 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d. 2025-01-08T10:46:23,827 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d. after waiting 0 ms 2025-01-08T10:46:23,827 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d. 2025-01-08T10:46:23,827 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] handler.UnassignRegionHandler(124): Close 9f13bcd219cfe423da66381b35d7e9b8 2025-01-08T10:46:23,827 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T10:46:23,827 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] regionserver.HRegion(1681): Closing 9f13bcd219cfe423da66381b35d7e9b8, disabling compactions & flushes 2025-01-08T10:46:23,827 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8. 2025-01-08T10:46:23,827 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8. 2025-01-08T10:46:23,827 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8. after waiting 0 ms 2025-01-08T10:46:23,828 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8. 2025-01-08T10:46:23,844 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/f2cb5b6ce155a147a1798fb975fd564d/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T10:46:23,845 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/9f13bcd219cfe423da66381b35d7e9b8/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T10:46:23,845 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:46:23,845 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:46:23,846 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d. 2025-01-08T10:46:23,846 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8. 2025-01-08T10:46:23,846 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] regionserver.HRegion(1635): Region close journal for f2cb5b6ce155a147a1798fb975fd564d: 2025-01-08T10:46:23,846 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] regionserver.HRegion(1635): Region close journal for 9f13bcd219cfe423da66381b35d7e9b8: 2025-01-08T10:46:23,847 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] handler.UnassignRegionHandler(170): Closed 9f13bcd219cfe423da66381b35d7e9b8 2025-01-08T10:46:23,848 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=76 updating hbase:meta row=9f13bcd219cfe423da66381b35d7e9b8, regionState=CLOSED 2025-01-08T10:46:23,848 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] handler.UnassignRegionHandler(170): Closed f2cb5b6ce155a147a1798fb975fd564d 2025-01-08T10:46:23,849 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=77 updating hbase:meta row=f2cb5b6ce155a147a1798fb975fd564d, regionState=CLOSED 2025-01-08T10:46:23,854 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=79, resume processing ppid=76 2025-01-08T10:46:23,854 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=79, ppid=76, state=SUCCESS; CloseRegionProcedure 9f13bcd219cfe423da66381b35d7e9b8, server=67f42ee69b4e,38323,1736333108259 in 176 msec 2025-01-08T10:46:23,855 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=78, resume processing ppid=77 2025-01-08T10:46:23,856 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=78, ppid=77, state=SUCCESS; CloseRegionProcedure f2cb5b6ce155a147a1798fb975fd564d, server=67f42ee69b4e,45003,1736333108039 in 178 msec 2025-01-08T10:46:23,856 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=76, ppid=75, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=9f13bcd219cfe423da66381b35d7e9b8, UNASSIGN in 184 msec 2025-01-08T10:46:23,857 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=77, resume processing ppid=75 2025-01-08T10:46:23,857 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=77, ppid=75, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=f2cb5b6ce155a147a1798fb975fd564d, UNASSIGN in 185 msec 2025-01-08T10:46:23,859 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=75, resume processing ppid=74 2025-01-08T10:46:23,859 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=75, ppid=74, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testExportFileSystemState in 190 msec 2025-01-08T10:46:23,861 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333183860"}]},"ts":"1736333183860"} 2025-01-08T10:46:23,862 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemState, state=DISABLED in hbase:meta 2025-01-08T10:46:23,875 INFO [PEWorker-5 {}] procedure.DisableTableProcedure(296): Set testtb-testExportFileSystemState to state=DISABLED 2025-01-08T10:46:23,877 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=74, state=SUCCESS; DisableTableProcedure table=testtb-testExportFileSystemState in 245 msec 2025-01-08T10:46:23,941 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=74 2025-01-08T10:46:23,942 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testExportFileSystemState, procId: 74 completed 2025-01-08T10:46:23,942 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.2 delete testtb-testExportFileSystemState 2025-01-08T10:46:23,943 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=80, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testExportFileSystemState 2025-01-08T10:46:23,944 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=80, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemState 2025-01-08T10:46:23,944 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testExportFileSystemState 2025-01-08T10:46:23,945 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=80, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemState 2025-01-08T10:46:23,947 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36283 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testExportFileSystemState 2025-01-08T10:46:23,949 DEBUG [HFileArchiver-7 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/9f13bcd219cfe423da66381b35d7e9b8 2025-01-08T10:46:23,949 DEBUG [HFileArchiver-8 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/f2cb5b6ce155a147a1798fb975fd564d 2025-01-08T10:46:23,951 DEBUG [HFileArchiver-7 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/9f13bcd219cfe423da66381b35d7e9b8/cf, FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/9f13bcd219cfe423da66381b35d7e9b8/recovered.edits] 2025-01-08T10:46:23,951 DEBUG [HFileArchiver-8 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/f2cb5b6ce155a147a1798fb975fd564d/cf, FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/f2cb5b6ce155a147a1798fb975fd564d/recovered.edits] 2025-01-08T10:46:23,956 DEBUG [HFileArchiver-8 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/f2cb5b6ce155a147a1798fb975fd564d/cf/b50639db54bf4cffa375e391dca52f5c to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportFileSystemState/f2cb5b6ce155a147a1798fb975fd564d/cf/b50639db54bf4cffa375e391dca52f5c 2025-01-08T10:46:23,956 DEBUG [HFileArchiver-7 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/9f13bcd219cfe423da66381b35d7e9b8/cf/e9d58b0004864757b4fde3925177d46d to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportFileSystemState/9f13bcd219cfe423da66381b35d7e9b8/cf/e9d58b0004864757b4fde3925177d46d 2025-01-08T10:46:23,959 DEBUG [HFileArchiver-8 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/f2cb5b6ce155a147a1798fb975fd564d/recovered.edits/9.seqid to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportFileSystemState/f2cb5b6ce155a147a1798fb975fd564d/recovered.edits/9.seqid 2025-01-08T10:46:23,960 DEBUG [HFileArchiver-8 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/f2cb5b6ce155a147a1798fb975fd564d 2025-01-08T10:46:23,962 DEBUG [HFileArchiver-7 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/9f13bcd219cfe423da66381b35d7e9b8/recovered.edits/9.seqid to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportFileSystemState/9f13bcd219cfe423da66381b35d7e9b8/recovered.edits/9.seqid 2025-01-08T10:46:23,962 DEBUG [HFileArchiver-7 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemState/9f13bcd219cfe423da66381b35d7e9b8 2025-01-08T10:46:23,963 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(313): Archived testtb-testExportFileSystemState regions 2025-01-08T10:46:23,963 DEBUG [PEWorker-2 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d 2025-01-08T10:46:23,964 DEBUG [PEWorker-2 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf] 2025-01-08T10:46:23,968 DEBUG [PEWorker-2 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/c4ca4238a0b923820dcc509a6f75849b202501085f987b4e97ca4ffe8e71a63b1de64287_f2cb5b6ce155a147a1798fb975fd564d to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/c4ca4238a0b923820dcc509a6f75849b202501085f987b4e97ca4ffe8e71a63b1de64287_f2cb5b6ce155a147a1798fb975fd564d 2025-01-08T10:46:23,970 DEBUG [PEWorker-2 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/d41d8cd98f00b204e9800998ecf8427e202501087e69e94c8e224ddcaab3c0b8e50fb119_9f13bcd219cfe423da66381b35d7e9b8 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/d41d8cd98f00b204e9800998ecf8427e202501087e69e94c8e224ddcaab3c0b8e50fb119_9f13bcd219cfe423da66381b35d7e9b8 2025-01-08T10:46:23,971 DEBUG [PEWorker-2 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d 2025-01-08T10:46:23,973 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=80, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemState 2025-01-08T10:46:23,975 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemState 2025-01-08T10:46:23,975 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemState 2025-01-08T10:46:23,975 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemState 2025-01-08T10:46:23,975 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemState 2025-01-08T10:46:23,976 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemState with data PBUF 2025-01-08T10:46:23,976 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemState with data PBUF 2025-01-08T10:46:23,976 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemState with data PBUF 2025-01-08T10:46:23,976 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemState with data PBUF 2025-01-08T10:46:23,976 WARN [PEWorker-2 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testtb-testExportFileSystemState from hbase:meta 2025-01-08T10:46:23,978 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testExportFileSystemState' descriptor. 2025-01-08T10:46:23,979 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=80, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemState 2025-01-08T10:46:23,979 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testExportFileSystemState' from region states. 2025-01-08T10:46:23,979 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736333183979"}]},"ts":"9223372036854775807"} 2025-01-08T10:46:23,980 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736333183979"}]},"ts":"9223372036854775807"} 2025-01-08T10:46:23,984 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T10:46:23,984 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => 9f13bcd219cfe423da66381b35d7e9b8, NAME => 'testtb-testExportFileSystemState,,1736333165891.9f13bcd219cfe423da66381b35d7e9b8.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => f2cb5b6ce155a147a1798fb975fd564d, NAME => 'testtb-testExportFileSystemState,1,1736333165891.f2cb5b6ce155a147a1798fb975fd564d.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T10:46:23,984 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testExportFileSystemState' as deleted. 2025-01-08T10:46:23,984 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736333183984"}]},"ts":"9223372036854775807"} 2025-01-08T10:46:23,986 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testExportFileSystemState state from META 2025-01-08T10:46:23,990 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemState 2025-01-08T10:46:23,990 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemState 2025-01-08T10:46:23,991 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemState 2025-01-08T10:46:23,991 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:46:23,991 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemState 2025-01-08T10:46:23,991 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:46:23,991 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:46:23,991 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:46:23,992 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=80 2025-01-08T10:46:24,006 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(133): Finished pid=80, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemState 2025-01-08T10:46:24,007 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=80, state=SUCCESS; DeleteTableProcedure table=testtb-testExportFileSystemState in 65 msec 2025-01-08T10:46:24,093 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=80 2025-01-08T10:46:24,093 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testExportFileSystemState, procId: 80 completed 2025-01-08T10:46:24,105 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.2 delete name: "emptySnaptb0-testExportFileSystemState" 2025-01-08T10:46:24,107 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(380): Deleting snapshot: emptySnaptb0-testExportFileSystemState 2025-01-08T10:46:24,109 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.2 delete name: "snaptb0-testExportFileSystemState" 2025-01-08T10:46:24,110 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testExportFileSystemState 2025-01-08T10:46:24,134 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: snapshot.TestMobSecureExportSnapshot#testExportFileSystemState Thread=785 (was 778) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_278065467_1 at /127.0.0.1:40680 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:37407 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-6-4 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:62) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:883) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-7 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (175901922) connection to localhost/127.0.0.1:37407 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: HFileArchiver-8 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-832679797_22 at /127.0.0.1:40696 [Waiting for operation #3] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #3 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LogDeleter #1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1177) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_278065467_1 at /127.0.0.1:60654 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-832679797_22 at /127.0.0.1:38740 [Waiting for operation #6] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: region-location-4 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-20 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-19 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-832679797_22 at /127.0.0.1:60672 [Waiting for operation #3] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-17 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Thread-2927 java.base@17.0.11/java.io.FileInputStream.readBytes(Native Method) java.base@17.0.11/java.io.FileInputStream.read(FileInputStream.java:276) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:282) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) java.base@17.0.11/sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:281) java.base@17.0.11/sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:324) java.base@17.0.11/sun.nio.cs.StreamDecoder.read(StreamDecoder.java:189) java.base@17.0.11/java.io.InputStreamReader.read(InputStreamReader.java:177) java.base@17.0.11/java.io.BufferedReader.fill(BufferedReader.java:162) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:329) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:396) app//org.apache.hadoop.util.Shell$1.run(Shell.java:1025) Potentially hanging thread: process reaper (pid 11892) java.base@17.0.11/java.lang.ProcessHandleImpl.waitForProcessExit0(Native Method) java.base@17.0.11/java.lang.ProcessHandleImpl$1.run(ProcessHandleImpl.java:150) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-18 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #4 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) - Thread LEAK? -, OpenFileDescriptor=803 (was 805), MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=657 (was 710), ProcessCount=18 (was 18), AvailableMemoryMB=2011 (was 2137) 2025-01-08T10:46:24,135 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=785 is superior to 500 2025-01-08T10:46:24,153 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: snapshot.TestMobSecureExportSnapshot#testConsecutiveExports Thread=785, OpenFileDescriptor=803, MaxFileDescriptor=1048576, SystemLoadAverage=657, ProcessCount=18, AvailableMemoryMB=2025 2025-01-08T10:46:24,154 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=785 is superior to 500 2025-01-08T10:46:24,157 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.2 create 'testtb-testConsecutiveExports', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T10:46:24,159 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=81, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testConsecutiveExports 2025-01-08T10:46:24,160 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=81, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testConsecutiveExports execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T10:46:24,161 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "testtb-testConsecutiveExports" procId is: 81 2025-01-08T10:46:24,161 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=81, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testConsecutiveExports execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T10:46:24,162 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=81 2025-01-08T10:46:24,174 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742035_1211 (size=440) 2025-01-08T10:46:24,174 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742035_1211 (size=440) 2025-01-08T10:46:24,174 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742035_1211 (size=440) 2025-01-08T10:46:24,177 INFO [RegionOpenAndInit-testtb-testConsecutiveExports-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => df917187823483f4cb65dabe50a92ccc, NAME => 'testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testtb-testConsecutiveExports', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:46:24,178 INFO [RegionOpenAndInit-testtb-testConsecutiveExports-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => 677a448659d57404fc829eb39d04dfa9, NAME => 'testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testtb-testConsecutiveExports', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:46:24,194 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742036_1212 (size=65) 2025-01-08T10:46:24,195 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742037_1213 (size=65) 2025-01-08T10:46:24,195 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742036_1212 (size=65) 2025-01-08T10:46:24,195 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742037_1213 (size=65) 2025-01-08T10:46:24,197 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742036_1212 (size=65) 2025-01-08T10:46:24,197 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742037_1213 (size=65) 2025-01-08T10:46:24,197 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:46:24,197 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:46:24,197 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-1 {}] regionserver.HRegion(1681): Closing 677a448659d57404fc829eb39d04dfa9, disabling compactions & flushes 2025-01-08T10:46:24,197 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-0 {}] regionserver.HRegion(1681): Closing df917187823483f4cb65dabe50a92ccc, disabling compactions & flushes 2025-01-08T10:46:24,197 INFO [RegionOpenAndInit-testtb-testConsecutiveExports-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc. 2025-01-08T10:46:24,197 INFO [RegionOpenAndInit-testtb-testConsecutiveExports-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9. 2025-01-08T10:46:24,197 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc. 2025-01-08T10:46:24,197 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9. 2025-01-08T10:46:24,197 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc. after waiting 0 ms 2025-01-08T10:46:24,197 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc. 2025-01-08T10:46:24,197 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9. after waiting 0 ms 2025-01-08T10:46:24,197 INFO [RegionOpenAndInit-testtb-testConsecutiveExports-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc. 2025-01-08T10:46:24,197 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9. 2025-01-08T10:46:24,197 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-0 {}] regionserver.HRegion(1635): Region close journal for df917187823483f4cb65dabe50a92ccc: 2025-01-08T10:46:24,197 INFO [RegionOpenAndInit-testtb-testConsecutiveExports-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9. 2025-01-08T10:46:24,197 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-1 {}] regionserver.HRegion(1635): Region close journal for 677a448659d57404fc829eb39d04dfa9: 2025-01-08T10:46:24,198 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=81, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testConsecutiveExports execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T10:46:24,199 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc.","families":{"info":[{"qualifier":"regioninfo","vlen":64,"tag":[],"timestamp":"1736333184199"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736333184199"}]},"ts":"1736333184199"} 2025-01-08T10:46:24,199 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9.","families":{"info":[{"qualifier":"regioninfo","vlen":64,"tag":[],"timestamp":"1736333184199"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736333184199"}]},"ts":"1736333184199"} 2025-01-08T10:46:24,202 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T10:46:24,203 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=81, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testConsecutiveExports execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T10:46:24,203 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testConsecutiveExports","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333184203"}]},"ts":"1736333184203"} 2025-01-08T10:46:24,204 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testConsecutiveExports, state=ENABLING in hbase:meta 2025-01-08T10:46:24,237 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(202): Hosts are {67f42ee69b4e=0} racks are {/default-rack=0} 2025-01-08T10:46:24,239 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T10:46:24,239 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T10:46:24,239 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T10:46:24,239 INFO [PEWorker-4 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T10:46:24,239 INFO [PEWorker-4 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T10:46:24,239 INFO [PEWorker-4 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T10:46:24,239 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T10:46:24,240 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=82, ppid=81, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=df917187823483f4cb65dabe50a92ccc, ASSIGN}, {pid=83, ppid=81, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=677a448659d57404fc829eb39d04dfa9, ASSIGN}] 2025-01-08T10:46:24,241 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=83, ppid=81, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=677a448659d57404fc829eb39d04dfa9, ASSIGN 2025-01-08T10:46:24,241 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=82, ppid=81, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=df917187823483f4cb65dabe50a92ccc, ASSIGN 2025-01-08T10:46:24,242 INFO [PEWorker-1 {}] assignment.TransitRegionStateProcedure(264): Starting pid=83, ppid=81, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=677a448659d57404fc829eb39d04dfa9, ASSIGN; state=OFFLINE, location=67f42ee69b4e,45003,1736333108039; forceNewPlan=false, retain=false 2025-01-08T10:46:24,242 INFO [PEWorker-3 {}] assignment.TransitRegionStateProcedure(264): Starting pid=82, ppid=81, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=df917187823483f4cb65dabe50a92ccc, ASSIGN; state=OFFLINE, location=67f42ee69b4e,36283,1736333108166; forceNewPlan=false, retain=false 2025-01-08T10:46:24,263 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=81 2025-01-08T10:46:24,392 INFO [67f42ee69b4e:34551 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T10:46:24,392 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=83 updating hbase:meta row=677a448659d57404fc829eb39d04dfa9, regionState=OPENING, regionLocation=67f42ee69b4e,45003,1736333108039 2025-01-08T10:46:24,392 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=82 updating hbase:meta row=df917187823483f4cb65dabe50a92ccc, regionState=OPENING, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:46:24,394 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=84, ppid=83, state=RUNNABLE; OpenRegionProcedure 677a448659d57404fc829eb39d04dfa9, server=67f42ee69b4e,45003,1736333108039}] 2025-01-08T10:46:24,395 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=85, ppid=82, state=RUNNABLE; OpenRegionProcedure df917187823483f4cb65dabe50a92ccc, server=67f42ee69b4e,36283,1736333108166}] 2025-01-08T10:46:24,465 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=81 2025-01-08T10:46:24,547 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:46:24,549 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:46:24,550 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] handler.AssignRegionHandler(135): Open testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9. 2025-01-08T10:46:24,551 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(7285): Opening region: {ENCODED => 677a448659d57404fc829eb39d04dfa9, NAME => 'testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T10:46:24,551 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9. service=AccessControlService 2025-01-08T10:46:24,551 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:46:24,551 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testConsecutiveExports 677a448659d57404fc829eb39d04dfa9 2025-01-08T10:46:24,552 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(894): Instantiated testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:46:24,552 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(7327): checking encryption for 677a448659d57404fc829eb39d04dfa9 2025-01-08T10:46:24,552 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(7330): checking classloading for 677a448659d57404fc829eb39d04dfa9 2025-01-08T10:46:24,553 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] handler.AssignRegionHandler(135): Open testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc. 2025-01-08T10:46:24,553 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(7285): Opening region: {ENCODED => df917187823483f4cb65dabe50a92ccc, NAME => 'testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T10:46:24,553 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc. service=AccessControlService 2025-01-08T10:46:24,554 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:46:24,554 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testConsecutiveExports df917187823483f4cb65dabe50a92ccc 2025-01-08T10:46:24,554 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(894): Instantiated testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:46:24,554 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(7327): checking encryption for df917187823483f4cb65dabe50a92ccc 2025-01-08T10:46:24,554 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(7330): checking classloading for df917187823483f4cb65dabe50a92ccc 2025-01-08T10:46:24,555 INFO [StoreOpener-677a448659d57404fc829eb39d04dfa9-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 677a448659d57404fc829eb39d04dfa9 2025-01-08T10:46:24,555 INFO [StoreOpener-df917187823483f4cb65dabe50a92ccc-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region df917187823483f4cb65dabe50a92ccc 2025-01-08T10:46:24,556 INFO [StoreOpener-677a448659d57404fc829eb39d04dfa9-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 677a448659d57404fc829eb39d04dfa9 columnFamilyName cf 2025-01-08T10:46:24,557 DEBUG [StoreOpener-677a448659d57404fc829eb39d04dfa9-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:46:24,558 INFO [StoreOpener-df917187823483f4cb65dabe50a92ccc-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region df917187823483f4cb65dabe50a92ccc columnFamilyName cf 2025-01-08T10:46:24,558 INFO [StoreOpener-677a448659d57404fc829eb39d04dfa9-1 {}] regionserver.HStore(327): Store=677a448659d57404fc829eb39d04dfa9/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:46:24,558 DEBUG [StoreOpener-df917187823483f4cb65dabe50a92ccc-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:46:24,559 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/677a448659d57404fc829eb39d04dfa9 2025-01-08T10:46:24,559 INFO [StoreOpener-df917187823483f4cb65dabe50a92ccc-1 {}] regionserver.HStore(327): Store=df917187823483f4cb65dabe50a92ccc/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:46:24,560 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/677a448659d57404fc829eb39d04dfa9 2025-01-08T10:46:24,560 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/df917187823483f4cb65dabe50a92ccc 2025-01-08T10:46:24,560 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/df917187823483f4cb65dabe50a92ccc 2025-01-08T10:46:24,562 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(1085): writing seq id for 677a448659d57404fc829eb39d04dfa9 2025-01-08T10:46:24,564 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/677a448659d57404fc829eb39d04dfa9/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:46:24,565 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(1102): Opened 677a448659d57404fc829eb39d04dfa9; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=61152567, jitterRate=-0.08875574171543121}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T10:46:24,566 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(1001): Region open journal for 677a448659d57404fc829eb39d04dfa9: 2025-01-08T10:46:24,567 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9., pid=84, masterSystemTime=1736333184547 2025-01-08T10:46:24,568 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(1085): writing seq id for df917187823483f4cb65dabe50a92ccc 2025-01-08T10:46:24,568 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9. 2025-01-08T10:46:24,569 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] handler.AssignRegionHandler(164): Opened testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9. 2025-01-08T10:46:24,569 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=83 updating hbase:meta row=677a448659d57404fc829eb39d04dfa9, regionState=OPEN, openSeqNum=2, regionLocation=67f42ee69b4e,45003,1736333108039 2025-01-08T10:46:24,572 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=84, resume processing ppid=83 2025-01-08T10:46:24,572 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=84, ppid=83, state=SUCCESS; OpenRegionProcedure 677a448659d57404fc829eb39d04dfa9, server=67f42ee69b4e,45003,1736333108039 in 177 msec 2025-01-08T10:46:24,573 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=83, ppid=81, state=SUCCESS; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=677a448659d57404fc829eb39d04dfa9, ASSIGN in 333 msec 2025-01-08T10:46:24,576 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/df917187823483f4cb65dabe50a92ccc/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:46:24,577 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(1102): Opened df917187823483f4cb65dabe50a92ccc; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=75286273, jitterRate=0.12185288965702057}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T10:46:24,577 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(1001): Region open journal for df917187823483f4cb65dabe50a92ccc: 2025-01-08T10:46:24,578 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc., pid=85, masterSystemTime=1736333184549 2025-01-08T10:46:24,579 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc. 2025-01-08T10:46:24,579 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] handler.AssignRegionHandler(164): Opened testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc. 2025-01-08T10:46:24,580 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=82 updating hbase:meta row=df917187823483f4cb65dabe50a92ccc, regionState=OPEN, openSeqNum=2, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:46:24,582 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=85, resume processing ppid=82 2025-01-08T10:46:24,582 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=85, ppid=82, state=SUCCESS; OpenRegionProcedure df917187823483f4cb65dabe50a92ccc, server=67f42ee69b4e,36283,1736333108166 in 186 msec 2025-01-08T10:46:24,584 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=82, resume processing ppid=81 2025-01-08T10:46:24,584 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=82, ppid=81, state=SUCCESS; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=df917187823483f4cb65dabe50a92ccc, ASSIGN in 343 msec 2025-01-08T10:46:24,585 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=81, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testConsecutiveExports execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T10:46:24,585 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testConsecutiveExports","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333184585"}]},"ts":"1736333184585"} 2025-01-08T10:46:24,586 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testConsecutiveExports, state=ENABLED in hbase:meta 2025-01-08T10:46:24,599 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=81, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testConsecutiveExports execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T10:46:24,599 DEBUG [PEWorker-3 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testConsecutiveExports jenkins: RWXCA 2025-01-08T10:46:24,602 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36283 {}] access.PermissionStorage(611): Read acl: entry[testtb-testConsecutiveExports], kv [jenkins: RWXCA] 2025-01-08T10:46:24,614 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:46:24,614 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:46:24,614 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:46:24,614 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:46:24,629 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testConsecutiveExports with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testConsecutiveExports \x00 \x01 \x02 \x03 \x04 2025-01-08T10:46:24,629 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testConsecutiveExports with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testConsecutiveExports \x00 \x01 \x02 \x03 \x04 2025-01-08T10:46:24,629 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testConsecutiveExports with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testConsecutiveExports \x00 \x01 \x02 \x03 \x04 2025-01-08T10:46:24,630 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testConsecutiveExports with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testConsecutiveExports \x00 \x01 \x02 \x03 \x04 2025-01-08T10:46:24,631 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=81, state=SUCCESS; CreateTableProcedure table=testtb-testConsecutiveExports in 472 msec 2025-01-08T10:46:24,766 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=81 2025-01-08T10:46:24,767 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testConsecutiveExports, procId: 81 completed 2025-01-08T10:46:24,770 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testConsecutiveExports 2025-01-08T10:46:24,770 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc. 2025-01-08T10:46:24,770 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T10:46:24,785 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.2 snapshot request for:{ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } 2025-01-08T10:46:24,785 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736333184785 (current time:1736333184785). 2025-01-08T10:46:24,785 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T10:46:24,785 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot emptySnaptb0-testConsecutiveExports VERSION not specified, setting to 2 2025-01-08T10:46:24,785 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T10:46:24,787 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x1d4fcfaa to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@49acf75 2025-01-08T10:46:24,799 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7e69b580, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:46:24,801 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:46:24,802 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:48198, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:46:24,804 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x1d4fcfaa to 127.0.0.1:52367 2025-01-08T10:46:24,804 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:46:24,805 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x2040cccf to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@7d7d334e 2025-01-08T10:46:24,830 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4e0b4cd4, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:46:24,832 DEBUG [hconnection-0x33ceb99e-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:46:24,834 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:48208, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:46:24,837 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x2040cccf to 127.0.0.1:52367 2025-01-08T10:46:24,837 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:46:24,838 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] access.PermissionStorage(611): Read acl: entry[testtb-testConsecutiveExports], kv [jenkins: RWXCA] 2025-01-08T10:46:24,839 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T10:46:24,839 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=86, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } 2025-01-08T10:46:24,840 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 }, snapshot procedure id = 86 2025-01-08T10:46:24,840 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T10:46:24,841 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=86 2025-01-08T10:46:24,841 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T10:46:24,845 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T10:46:24,876 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742038_1214 (size=161) 2025-01-08T10:46:24,876 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742038_1214 (size=161) 2025-01-08T10:46:24,876 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742038_1214 (size=161) 2025-01-08T10:46:24,878 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T10:46:24,878 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=87, ppid=86, state=RUNNABLE; SnapshotRegionProcedure df917187823483f4cb65dabe50a92ccc}, {pid=88, ppid=86, state=RUNNABLE; SnapshotRegionProcedure 677a448659d57404fc829eb39d04dfa9}] 2025-01-08T10:46:24,879 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=87, ppid=86, state=RUNNABLE; SnapshotRegionProcedure df917187823483f4cb65dabe50a92ccc 2025-01-08T10:46:24,879 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=88, ppid=86, state=RUNNABLE; SnapshotRegionProcedure 677a448659d57404fc829eb39d04dfa9 2025-01-08T10:46:24,941 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=86 2025-01-08T10:46:25,030 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:46:25,030 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:46:25,031 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=36283 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=87 2025-01-08T10:46:25,031 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=87}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc. 2025-01-08T10:46:25,031 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=45003 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=88 2025-01-08T10:46:25,032 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=87}] regionserver.HRegion(2538): Flush status journal for df917187823483f4cb65dabe50a92ccc: 2025-01-08T10:46:25,032 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=88}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9. 2025-01-08T10:46:25,032 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=87}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc. for emptySnaptb0-testConsecutiveExports completed. 2025-01-08T10:46:25,032 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=88}] regionserver.HRegion(2538): Flush status journal for 677a448659d57404fc829eb39d04dfa9: 2025-01-08T10:46:25,032 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=88}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9. for emptySnaptb0-testConsecutiveExports completed. 2025-01-08T10:46:25,032 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=87}] snapshot.SnapshotManifest(241): Storing 'testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc.' region-info for snapshot=emptySnaptb0-testConsecutiveExports 2025-01-08T10:46:25,032 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=87}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:46:25,032 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=88}] snapshot.SnapshotManifest(241): Storing 'testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9.' region-info for snapshot=emptySnaptb0-testConsecutiveExports 2025-01-08T10:46:25,032 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=88}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:46:25,032 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=87}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T10:46:25,032 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=88}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T10:46:25,050 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742040_1216 (size=68) 2025-01-08T10:46:25,051 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742040_1216 (size=68) 2025-01-08T10:46:25,051 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742040_1216 (size=68) 2025-01-08T10:46:25,054 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=88}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9. 2025-01-08T10:46:25,055 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=88}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=88 2025-01-08T10:46:25,055 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=88 2025-01-08T10:46:25,055 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testConsecutiveExports on region 677a448659d57404fc829eb39d04dfa9 2025-01-08T10:46:25,055 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=88, ppid=86, state=RUNNABLE; SnapshotRegionProcedure 677a448659d57404fc829eb39d04dfa9 2025-01-08T10:46:25,056 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742039_1215 (size=68) 2025-01-08T10:46:25,057 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742039_1215 (size=68) 2025-01-08T10:46:25,057 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742039_1215 (size=68) 2025-01-08T10:46:25,058 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=87}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc. 2025-01-08T10:46:25,058 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=87}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=87 2025-01-08T10:46:25,058 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=88, ppid=86, state=SUCCESS; SnapshotRegionProcedure 677a448659d57404fc829eb39d04dfa9 in 178 msec 2025-01-08T10:46:25,058 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=87 2025-01-08T10:46:25,058 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testConsecutiveExports on region df917187823483f4cb65dabe50a92ccc 2025-01-08T10:46:25,059 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=87, ppid=86, state=RUNNABLE; SnapshotRegionProcedure df917187823483f4cb65dabe50a92ccc 2025-01-08T10:46:25,061 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=87, resume processing ppid=86 2025-01-08T10:46:25,061 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T10:46:25,062 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=87, ppid=86, state=SUCCESS; SnapshotRegionProcedure df917187823483f4cb65dabe50a92ccc in 181 msec 2025-01-08T10:46:25,062 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T10:46:25,063 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T10:46:25,064 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T10:46:25,064 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:46:25,064 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(328): No files under family: cf 2025-01-08T10:46:25,083 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742041_1217 (size=60) 2025-01-08T10:46:25,084 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742041_1217 (size=60) 2025-01-08T10:46:25,084 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742041_1217 (size=60) 2025-01-08T10:46:25,143 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=86 2025-01-08T10:46:25,444 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=86 2025-01-08T10:46:25,485 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T10:46:25,486 DEBUG [PEWorker-2 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for emptySnaptb0-testConsecutiveExports 2025-01-08T10:46:25,486 DEBUG [PEWorker-2 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/emptySnaptb0-testConsecutiveExports 2025-01-08T10:46:25,501 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742042_1218 (size=641) 2025-01-08T10:46:25,501 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742042_1218 (size=641) 2025-01-08T10:46:25,501 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742042_1218 (size=641) 2025-01-08T10:46:25,504 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T10:46:25,512 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T10:46:25,512 DEBUG [PEWorker-2 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/emptySnaptb0-testConsecutiveExports to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/emptySnaptb0-testConsecutiveExports 2025-01-08T10:46:25,519 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T10:46:25,519 DEBUG [PEWorker-2 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 }, snapshot procedure id = 86 2025-01-08T10:46:25,534 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=86, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } in 680 msec 2025-01-08T10:46:25,946 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=86 2025-01-08T10:46:25,946 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testConsecutiveExports, procId: 86 completed 2025-01-08T10:46:25,950 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36283 {}] regionserver.HRegion(8254): writing data to region testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T10:46:25,951 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=45003 {}] regionserver.HRegion(8254): writing data to region testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T10:46:25,955 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testConsecutiveExports 2025-01-08T10:46:25,955 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc. 2025-01-08T10:46:25,956 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T10:46:25,971 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.2 snapshot request for:{ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } 2025-01-08T10:46:25,972 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736333185971 (current time:1736333185971). 2025-01-08T10:46:25,972 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T10:46:25,972 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testConsecutiveExports VERSION not specified, setting to 2 2025-01-08T10:46:25,972 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T10:46:25,973 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x15441a27 to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@53005358 2025-01-08T10:46:25,992 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@37c63669, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:46:25,993 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:46:25,995 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:48212, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:46:25,997 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x15441a27 to 127.0.0.1:52367 2025-01-08T10:46:25,997 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:46:25,999 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x6fefa33e to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@3e24a708 2025-01-08T10:46:26,031 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@3857abdb, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:46:26,035 DEBUG [hconnection-0x1ac1972c-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:46:26,036 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:48214, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:46:26,040 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x6fefa33e to 127.0.0.1:52367 2025-01-08T10:46:26,040 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:46:26,040 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] access.PermissionStorage(611): Read acl: entry[testtb-testConsecutiveExports], kv [jenkins: RWXCA] 2025-01-08T10:46:26,042 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T10:46:26,043 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=89, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } 2025-01-08T10:46:26,043 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 }, snapshot procedure id = 89 2025-01-08T10:46:26,044 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T10:46:26,045 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=89 2025-01-08T10:46:26,045 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T10:46:26,048 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T10:46:26,072 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742043_1219 (size=156) 2025-01-08T10:46:26,072 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742043_1219 (size=156) 2025-01-08T10:46:26,076 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742043_1219 (size=156) 2025-01-08T10:46:26,077 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T10:46:26,078 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=90, ppid=89, state=RUNNABLE; SnapshotRegionProcedure df917187823483f4cb65dabe50a92ccc}, {pid=91, ppid=89, state=RUNNABLE; SnapshotRegionProcedure 677a448659d57404fc829eb39d04dfa9}] 2025-01-08T10:46:26,079 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=91, ppid=89, state=RUNNABLE; SnapshotRegionProcedure 677a448659d57404fc829eb39d04dfa9 2025-01-08T10:46:26,079 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=90, ppid=89, state=RUNNABLE; SnapshotRegionProcedure df917187823483f4cb65dabe50a92ccc 2025-01-08T10:46:26,146 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=89 2025-01-08T10:46:26,230 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:46:26,230 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:46:26,230 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=36283 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=90 2025-01-08T10:46:26,231 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=45003 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=91 2025-01-08T10:46:26,231 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc. 2025-01-08T10:46:26,231 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9. 2025-01-08T10:46:26,231 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.HRegion(2837): Flushing df917187823483f4cb65dabe50a92ccc 1/1 column families, dataSize=199 B heapSize=688 B 2025-01-08T10:46:26,231 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.HRegion(2837): Flushing 677a448659d57404fc829eb39d04dfa9 1/1 column families, dataSize=3.06 KB heapSize=6.86 KB 2025-01-08T10:46:26,251 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108037a5f0bd72a48acb4762718720e5abf_df917187823483f4cb65dabe50a92ccc is 71, key is 0ba57f4843de00f0cbc2a4c94408dcf3/cf:q/1736333185950/Put/seqid=0 2025-01-08T10:46:26,265 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b202501087c52c1f04b5d4722a0c956a8eccc0d94_677a448659d57404fc829eb39d04dfa9 is 71, key is 12a490695655ebfea95746244cb98ee4/cf:q/1736333185951/Put/seqid=0 2025-01-08T10:46:26,268 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742044_1220 (size=5102) 2025-01-08T10:46:26,268 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742044_1220 (size=5102) 2025-01-08T10:46:26,270 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742044_1220 (size=5102) 2025-01-08T10:46:26,270 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:46:26,276 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108037a5f0bd72a48acb4762718720e5abf_df917187823483f4cb65dabe50a92ccc to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/d41d8cd98f00b204e9800998ecf8427e20250108037a5f0bd72a48acb4762718720e5abf_df917187823483f4cb65dabe50a92ccc 2025-01-08T10:46:26,282 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742045_1221 (size=8171) 2025-01-08T10:46:26,283 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742045_1221 (size=8171) 2025-01-08T10:46:26,283 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742045_1221 (size=8171) 2025-01-08T10:46:26,284 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/df917187823483f4cb65dabe50a92ccc/.tmp/cf/6112f6e037fd410c84ce02256175994f, store: [table=testtb-testConsecutiveExports family=cf region=df917187823483f4cb65dabe50a92ccc] 2025-01-08T10:46:26,284 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:46:26,285 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/df917187823483f4cb65dabe50a92ccc/.tmp/cf/6112f6e037fd410c84ce02256175994f is 206, key is 0da2db3f2fda1ca5eccbbd6855fb1d882/cf:q/1736333185950/Put/seqid=0 2025-01-08T10:46:26,292 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b202501087c52c1f04b5d4722a0c956a8eccc0d94_677a448659d57404fc829eb39d04dfa9 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/c4ca4238a0b923820dcc509a6f75849b202501087c52c1f04b5d4722a0c956a8eccc0d94_677a448659d57404fc829eb39d04dfa9 2025-01-08T10:46:26,293 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/677a448659d57404fc829eb39d04dfa9/.tmp/cf/fc8a56c09f1a486fbad516fee9a400d4, store: [table=testtb-testConsecutiveExports family=cf region=677a448659d57404fc829eb39d04dfa9] 2025-01-08T10:46:26,293 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/677a448659d57404fc829eb39d04dfa9/.tmp/cf/fc8a56c09f1a486fbad516fee9a400d4 is 206, key is 1e7f13e3e54e2d0049b11d29bc010e3a1/cf:q/1736333185951/Put/seqid=0 2025-01-08T10:46:26,303 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742046_1222 (size=5906) 2025-01-08T10:46:26,304 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742046_1222 (size=5906) 2025-01-08T10:46:26,304 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742046_1222 (size=5906) 2025-01-08T10:46:26,304 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=199, hasBloomFilter=true, into tmp file hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/df917187823483f4cb65dabe50a92ccc/.tmp/cf/6112f6e037fd410c84ce02256175994f 2025-01-08T10:46:26,308 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742047_1223 (size=14853) 2025-01-08T10:46:26,308 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742047_1223 (size=14853) 2025-01-08T10:46:26,309 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742047_1223 (size=14853) 2025-01-08T10:46:26,309 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=3.1 K, hasBloomFilter=true, into tmp file hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/677a448659d57404fc829eb39d04dfa9/.tmp/cf/fc8a56c09f1a486fbad516fee9a400d4 2025-01-08T10:46:26,312 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/df917187823483f4cb65dabe50a92ccc/.tmp/cf/6112f6e037fd410c84ce02256175994f as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/df917187823483f4cb65dabe50a92ccc/cf/6112f6e037fd410c84ce02256175994f 2025-01-08T10:46:26,315 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/677a448659d57404fc829eb39d04dfa9/.tmp/cf/fc8a56c09f1a486fbad516fee9a400d4 as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/677a448659d57404fc829eb39d04dfa9/cf/fc8a56c09f1a486fbad516fee9a400d4 2025-01-08T10:46:26,321 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/df917187823483f4cb65dabe50a92ccc/cf/6112f6e037fd410c84ce02256175994f, entries=3, sequenceid=6, filesize=5.8 K 2025-01-08T10:46:26,322 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.HRegion(3040): Finished flush of dataSize ~199 B/199, heapSize ~672 B/672, currentSize=0 B/0 for df917187823483f4cb65dabe50a92ccc in 91ms, sequenceid=6, compaction requested=false 2025-01-08T10:46:26,322 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testConsecutiveExports' 2025-01-08T10:46:26,323 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.HRegion(2538): Flush status journal for df917187823483f4cb65dabe50a92ccc: 2025-01-08T10:46:26,323 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc. for snaptb0-testConsecutiveExports completed. 2025-01-08T10:46:26,323 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] snapshot.SnapshotManifest(241): Storing 'testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc.' region-info for snapshot=snaptb0-testConsecutiveExports 2025-01-08T10:46:26,323 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:46:26,324 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/df917187823483f4cb65dabe50a92ccc/cf/6112f6e037fd410c84ce02256175994f] hfiles 2025-01-08T10:46:26,324 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/df917187823483f4cb65dabe50a92ccc/cf/6112f6e037fd410c84ce02256175994f for snapshot=snaptb0-testConsecutiveExports 2025-01-08T10:46:26,324 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/677a448659d57404fc829eb39d04dfa9/cf/fc8a56c09f1a486fbad516fee9a400d4, entries=47, sequenceid=6, filesize=14.5 K 2025-01-08T10:46:26,324 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.HRegion(3040): Finished flush of dataSize ~3.06 KB/3137, heapSize ~6.84 KB/7008, currentSize=0 B/0 for 677a448659d57404fc829eb39d04dfa9 in 93ms, sequenceid=6, compaction requested=false 2025-01-08T10:46:26,325 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.HRegion(2538): Flush status journal for 677a448659d57404fc829eb39d04dfa9: 2025-01-08T10:46:26,325 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9. for snaptb0-testConsecutiveExports completed. 2025-01-08T10:46:26,325 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] snapshot.SnapshotManifest(241): Storing 'testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9.' region-info for snapshot=snaptb0-testConsecutiveExports 2025-01-08T10:46:26,325 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:46:26,325 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/677a448659d57404fc829eb39d04dfa9/cf/fc8a56c09f1a486fbad516fee9a400d4] hfiles 2025-01-08T10:46:26,325 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/677a448659d57404fc829eb39d04dfa9/cf/fc8a56c09f1a486fbad516fee9a400d4 for snapshot=snaptb0-testConsecutiveExports 2025-01-08T10:46:26,343 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742048_1224 (size=107) 2025-01-08T10:46:26,343 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742048_1224 (size=107) 2025-01-08T10:46:26,343 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742048_1224 (size=107) 2025-01-08T10:46:26,344 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc. 2025-01-08T10:46:26,344 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=90 2025-01-08T10:46:26,345 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=90 2025-01-08T10:46:26,345 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testConsecutiveExports on region df917187823483f4cb65dabe50a92ccc 2025-01-08T10:46:26,345 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=90, ppid=89, state=RUNNABLE; SnapshotRegionProcedure df917187823483f4cb65dabe50a92ccc 2025-01-08T10:46:26,347 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=90, ppid=89, state=SUCCESS; SnapshotRegionProcedure df917187823483f4cb65dabe50a92ccc in 268 msec 2025-01-08T10:46:26,348 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=89 2025-01-08T10:46:26,352 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742049_1225 (size=107) 2025-01-08T10:46:26,352 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742049_1225 (size=107) 2025-01-08T10:46:26,352 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742049_1225 (size=107) 2025-01-08T10:46:26,353 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9. 2025-01-08T10:46:26,353 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=91 2025-01-08T10:46:26,354 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=91 2025-01-08T10:46:26,354 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testConsecutiveExports on region 677a448659d57404fc829eb39d04dfa9 2025-01-08T10:46:26,354 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=91, ppid=89, state=RUNNABLE; SnapshotRegionProcedure 677a448659d57404fc829eb39d04dfa9 2025-01-08T10:46:26,356 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=91, resume processing ppid=89 2025-01-08T10:46:26,357 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=91, ppid=89, state=SUCCESS; SnapshotRegionProcedure 677a448659d57404fc829eb39d04dfa9 in 277 msec 2025-01-08T10:46:26,357 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T10:46:26,357 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T10:46:26,358 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T10:46:26,358 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T10:46:26,358 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:46:26,360 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/c4ca4238a0b923820dcc509a6f75849b202501087c52c1f04b5d4722a0c956a8eccc0d94_677a448659d57404fc829eb39d04dfa9, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/d41d8cd98f00b204e9800998ecf8427e20250108037a5f0bd72a48acb4762718720e5abf_df917187823483f4cb65dabe50a92ccc] hfiles 2025-01-08T10:46:26,360 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/c4ca4238a0b923820dcc509a6f75849b202501087c52c1f04b5d4722a0c956a8eccc0d94_677a448659d57404fc829eb39d04dfa9 2025-01-08T10:46:26,360 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/d41d8cd98f00b204e9800998ecf8427e20250108037a5f0bd72a48acb4762718720e5abf_df917187823483f4cb65dabe50a92ccc 2025-01-08T10:46:26,370 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742050_1226 (size=291) 2025-01-08T10:46:26,371 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742050_1226 (size=291) 2025-01-08T10:46:26,371 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742050_1226 (size=291) 2025-01-08T10:46:26,373 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T10:46:26,373 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testConsecutiveExports 2025-01-08T10:46:26,374 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/snaptb0-testConsecutiveExports 2025-01-08T10:46:26,396 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742051_1227 (size=951) 2025-01-08T10:46:26,396 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742051_1227 (size=951) 2025-01-08T10:46:26,397 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742051_1227 (size=951) 2025-01-08T10:46:26,400 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T10:46:26,408 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T10:46:26,409 DEBUG [PEWorker-1 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/snaptb0-testConsecutiveExports to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testConsecutiveExports 2025-01-08T10:46:26,410 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T10:46:26,410 DEBUG [PEWorker-1 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 }, snapshot procedure id = 89 2025-01-08T10:46:26,411 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=89, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } in 368 msec 2025-01-08T10:46:26,649 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=89 2025-01-08T10:46:26,649 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testConsecutiveExports, procId: 89 completed 2025-01-08T10:46:26,649 INFO [Time-limited test {}] snapshot.TestExportSnapshot(476): Local export destination path: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333186649 2025-01-08T10:46:26,649 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=file:///, tgtDir=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333186649, rawTgtDir=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333186649, srcFsUri=hdfs://localhost:37495, srcDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:46:26,684 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:37495, inputRoot=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:46:26,684 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=org.apache.hadoop.fs.LocalFileSystem@5d4621f9, outputRoot=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333186649, skipTmp=false, initialOutputSnapshotDir=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333186649/.hbase-snapshot/.tmp/snaptb0-testConsecutiveExports 2025-01-08T10:46:26,686 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T10:46:26,692 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testConsecutiveExports to file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333186649/.hbase-snapshot/.tmp/snaptb0-testConsecutiveExports 2025-01-08T10:46:26,912 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-6321259572578634042.jar 2025-01-08T10:46:26,913 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:26,913 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:26,913 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:27,596 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testConsecutiveExports 2025-01-08T10:46:27,596 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testConsecutiveExports Metrics about Tables on a single HBase RegionServer 2025-01-08T10:46:27,597 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemState 2025-01-08T10:46:27,856 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0003_000001 (auth:SIMPLE) from 127.0.0.1:36112 2025-01-08T10:46:27,866 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_2/usercache/jenkins/appcache/application_1736333116683_0003/container_1736333116683_0003_01_000001/launch_container.sh] 2025-01-08T10:46:27,866 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_2/usercache/jenkins/appcache/application_1736333116683_0003/container_1736333116683_0003_01_000001/container_tokens] 2025-01-08T10:46:27,866 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_2/usercache/jenkins/appcache/application_1736333116683_0003/container_1736333116683_0003_01_000001/sysfs] 2025-01-08T10:46:27,941 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-8293776099013943472.jar 2025-01-08T10:46:27,942 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:27,942 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:28,011 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-10301777667951686373.jar 2025-01-08T10:46:28,011 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:28,011 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:28,012 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:28,012 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:28,012 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:28,012 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:28,013 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T10:46:28,013 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T10:46:28,013 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T10:46:28,013 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T10:46:28,013 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T10:46:28,014 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T10:46:28,014 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T10:46:28,014 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T10:46:28,014 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T10:46:28,015 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T10:46:28,015 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T10:46:28,015 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T10:46:28,015 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:46:28,016 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:46:28,016 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:46:28,016 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:46:28,016 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:46:28,017 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:46:28,017 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:46:28,070 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742052_1228 (size=127628) 2025-01-08T10:46:28,071 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742052_1228 (size=127628) 2025-01-08T10:46:28,071 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742052_1228 (size=127628) 2025-01-08T10:46:28,085 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742053_1229 (size=2172137) 2025-01-08T10:46:28,085 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742053_1229 (size=2172137) 2025-01-08T10:46:28,086 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742053_1229 (size=2172137) 2025-01-08T10:46:28,494 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742054_1230 (size=213228) 2025-01-08T10:46:28,494 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742054_1230 (size=213228) 2025-01-08T10:46:28,495 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742054_1230 (size=213228) 2025-01-08T10:46:28,507 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742055_1231 (size=1877034) 2025-01-08T10:46:28,507 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742055_1231 (size=1877034) 2025-01-08T10:46:28,508 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742055_1231 (size=1877034) 2025-01-08T10:46:28,525 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742056_1232 (size=451756) 2025-01-08T10:46:28,527 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742056_1232 (size=451756) 2025-01-08T10:46:28,527 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742056_1232 (size=451756) 2025-01-08T10:46:28,549 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742057_1233 (size=533455) 2025-01-08T10:46:28,550 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742057_1233 (size=533455) 2025-01-08T10:46:28,553 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742057_1233 (size=533455) 2025-01-08T10:46:28,584 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742058_1234 (size=7280644) 2025-01-08T10:46:28,585 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742058_1234 (size=7280644) 2025-01-08T10:46:28,585 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742058_1234 (size=7280644) 2025-01-08T10:46:28,610 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742059_1235 (size=4188619) 2025-01-08T10:46:28,610 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742059_1235 (size=4188619) 2025-01-08T10:46:28,611 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742059_1235 (size=4188619) 2025-01-08T10:46:29,016 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742060_1236 (size=20406) 2025-01-08T10:46:29,016 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742060_1236 (size=20406) 2025-01-08T10:46:29,017 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742060_1236 (size=20406) 2025-01-08T10:46:29,024 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742061_1237 (size=75495) 2025-01-08T10:46:29,024 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742061_1237 (size=75495) 2025-01-08T10:46:29,024 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742061_1237 (size=75495) 2025-01-08T10:46:29,031 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742062_1238 (size=45609) 2025-01-08T10:46:29,032 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742062_1238 (size=45609) 2025-01-08T10:46:29,033 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742062_1238 (size=45609) 2025-01-08T10:46:29,040 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742063_1239 (size=110084) 2025-01-08T10:46:29,040 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742063_1239 (size=110084) 2025-01-08T10:46:29,041 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742063_1239 (size=110084) 2025-01-08T10:46:29,050 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742064_1240 (size=1323991) 2025-01-08T10:46:29,050 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742064_1240 (size=1323991) 2025-01-08T10:46:29,051 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742064_1240 (size=1323991) 2025-01-08T10:46:29,057 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742065_1241 (size=23076) 2025-01-08T10:46:29,058 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742065_1241 (size=23076) 2025-01-08T10:46:29,058 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742065_1241 (size=23076) 2025-01-08T10:46:29,067 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742066_1242 (size=126803) 2025-01-08T10:46:29,067 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742066_1242 (size=126803) 2025-01-08T10:46:29,067 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742066_1242 (size=126803) 2025-01-08T10:46:29,374 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T10:46:29,475 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742067_1243 (size=322274) 2025-01-08T10:46:29,475 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742067_1243 (size=322274) 2025-01-08T10:46:29,475 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742067_1243 (size=322274) 2025-01-08T10:46:29,486 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742068_1244 (size=1832290) 2025-01-08T10:46:29,487 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742068_1244 (size=1832290) 2025-01-08T10:46:29,487 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742068_1244 (size=1832290) 2025-01-08T10:46:29,494 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742069_1245 (size=30081) 2025-01-08T10:46:29,494 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742069_1245 (size=30081) 2025-01-08T10:46:29,494 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742069_1245 (size=30081) 2025-01-08T10:46:29,500 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742070_1246 (size=53616) 2025-01-08T10:46:29,501 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742070_1246 (size=53616) 2025-01-08T10:46:29,501 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742070_1246 (size=53616) 2025-01-08T10:46:29,507 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742071_1247 (size=29229) 2025-01-08T10:46:29,508 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742071_1247 (size=29229) 2025-01-08T10:46:29,508 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742071_1247 (size=29229) 2025-01-08T10:46:29,514 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742072_1248 (size=169089) 2025-01-08T10:46:29,515 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742072_1248 (size=169089) 2025-01-08T10:46:29,515 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742072_1248 (size=169089) 2025-01-08T10:46:29,526 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742073_1249 (size=912094) 2025-01-08T10:46:29,526 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742073_1249 (size=912094) 2025-01-08T10:46:29,526 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742073_1249 (size=912094) 2025-01-08T10:46:29,561 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742074_1250 (size=6350714) 2025-01-08T10:46:29,561 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742074_1250 (size=6350714) 2025-01-08T10:46:29,562 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742074_1250 (size=6350714) 2025-01-08T10:46:29,581 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742075_1251 (size=5175431) 2025-01-08T10:46:29,581 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742075_1251 (size=5175431) 2025-01-08T10:46:29,581 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742075_1251 (size=5175431) 2025-01-08T10:46:29,589 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742076_1252 (size=136454) 2025-01-08T10:46:29,589 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742076_1252 (size=136454) 2025-01-08T10:46:29,590 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742076_1252 (size=136454) 2025-01-08T10:46:29,606 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742077_1253 (size=3317408) 2025-01-08T10:46:29,606 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742077_1253 (size=3317408) 2025-01-08T10:46:29,606 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742077_1253 (size=3317408) 2025-01-08T10:46:29,628 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742078_1254 (size=503880) 2025-01-08T10:46:29,628 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742078_1254 (size=503880) 2025-01-08T10:46:29,628 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742078_1254 (size=503880) 2025-01-08T10:46:29,648 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742079_1255 (size=4695811) 2025-01-08T10:46:29,648 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742079_1255 (size=4695811) 2025-01-08T10:46:29,648 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742079_1255 (size=4695811) 2025-01-08T10:46:29,650 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T10:46:29,652 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'snaptb0-testConsecutiveExports' hfile list 2025-01-08T10:46:29,654 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(751): export split=0 size=33.2 K 2025-01-08T10:46:29,664 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742080_1256 (size=714) 2025-01-08T10:46:29,664 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742080_1256 (size=714) 2025-01-08T10:46:29,664 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742080_1256 (size=714) 2025-01-08T10:46:29,672 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742081_1257 (size=15) 2025-01-08T10:46:29,672 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742081_1257 (size=15) 2025-01-08T10:46:29,672 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742081_1257 (size=15) 2025-01-08T10:46:29,687 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742082_1258 (size=304976) 2025-01-08T10:46:29,688 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742082_1258 (size=304976) 2025-01-08T10:46:29,688 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742082_1258 (size=304976) 2025-01-08T10:46:29,706 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T10:46:29,706 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T10:46:29,827 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0004_000001 (auth:SIMPLE) from 127.0.0.1:35064 2025-01-08T10:46:31,924 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xffcea3966f5fb757 with lease ID 0xe506f1aed7680301: from storage DS-96de2396-f7dd-4dd7-805c-c34c70b41204 node DatanodeRegistration(127.0.0.1:43487, datanodeUuid=2039b132-af90-4efa-b0f3-2607003b309d, infoPort=34377, infoSecurePort=0, ipcPort=45989, storageInfo=lv=-57;cid=testClusterID;nsid=1138699459;c=1736333101274), blocks: 70, hasStaleStorage: false, processing time: 1 msecs, invalidatedBlocks: 25 2025-01-08T10:46:31,925 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xffcea3966f5fb757 with lease ID 0xe506f1aed7680301: from storage DS-3a6862aa-5072-4346-a310-a3fc8b9fc69c node DatanodeRegistration(127.0.0.1:43487, datanodeUuid=2039b132-af90-4efa-b0f3-2607003b309d, infoPort=34377, infoSecurePort=0, ipcPort=45989, storageInfo=lv=-57;cid=testClusterID;nsid=1138699459;c=1736333101274), blocks: 68, hasStaleStorage: false, processing time: 1 msecs, invalidatedBlocks: 25 2025-01-08T10:46:35,532 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0004_000001 (auth:SIMPLE) from 127.0.0.1:39808 2025-01-08T10:46:35,713 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T10:46:35,786 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742083_1259 (size=350650) 2025-01-08T10:46:35,786 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742083_1259 (size=350650) 2025-01-08T10:46:35,786 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742083_1259 (size=350650) 2025-01-08T10:46:37,813 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0004_000001 (auth:SIMPLE) from 127.0.0.1:54214 2025-01-08T10:46:41,994 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742084_1260 (size=17451) 2025-01-08T10:46:41,994 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742084_1260 (size=17451) 2025-01-08T10:46:41,995 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742084_1260 (size=17451) 2025-01-08T10:46:42,004 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742085_1261 (size=462) 2025-01-08T10:46:42,004 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742085_1261 (size=462) 2025-01-08T10:46:42,004 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742085_1261 (size=462) 2025-01-08T10:46:42,041 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742086_1262 (size=17451) 2025-01-08T10:46:42,042 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742086_1262 (size=17451) 2025-01-08T10:46:42,042 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742086_1262 (size=17451) 2025-01-08T10:46:42,075 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742087_1263 (size=350650) 2025-01-08T10:46:42,075 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742087_1263 (size=350650) 2025-01-08T10:46:42,075 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742087_1263 (size=350650) 2025-01-08T10:46:42,107 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0004_000001 (auth:SIMPLE) from 127.0.0.1:54228 2025-01-08T10:46:43,900 INFO [Time-limited test {}] snapshot.ExportSnapshot(1207): Finalize the Snapshot Export 2025-01-08T10:46:43,901 INFO [Time-limited test {}] snapshot.ExportSnapshot(1218): Verify the exported snapshot's expiration status and integrity. 2025-01-08T10:46:43,905 INFO [Time-limited test {}] snapshot.ExportSnapshot(1224): Export Completed: snaptb0-testConsecutiveExports 2025-01-08T10:46:43,905 INFO [Time-limited test {}] snapshot.TestExportSnapshot(362): Exported snapshot 2025-01-08T10:46:43,905 INFO [Time-limited test {}] snapshot.TestExportSnapshot(373): Verified filesystem state 2025-01-08T10:46:43,905 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-832679797_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testConsecutiveExports at hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testConsecutiveExports 2025-01-08T10:46:43,906 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testConsecutiveExports/.snapshotinfo 2025-01-08T10:46:43,906 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testConsecutiveExports/data.manifest 2025-01-08T10:46:43,906 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in org.apache.hadoop.fs.LocalFileSystem@5d4621f9 in root file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333186649/.hbase-snapshot/snaptb0-testConsecutiveExports at file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333186649/.hbase-snapshot/snaptb0-testConsecutiveExports 2025-01-08T10:46:43,906 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333186649/.hbase-snapshot/snaptb0-testConsecutiveExports/data.manifest 2025-01-08T10:46:43,906 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333186649/.hbase-snapshot/snaptb0-testConsecutiveExports/.snapshotinfo 2025-01-08T10:46:43,909 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=file:///, tgtDir=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333186649, rawTgtDir=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333186649, srcFsUri=hdfs://localhost:37495, srcDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:46:43,951 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:37495, inputRoot=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:46:43,951 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=org.apache.hadoop.fs.LocalFileSystem@5d4621f9, outputRoot=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333186649, skipTmp=false, initialOutputSnapshotDir=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333186649/.hbase-snapshot/.tmp/snaptb0-testConsecutiveExports 2025-01-08T10:46:43,954 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T10:46:43,963 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testConsecutiveExports to file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333186649/.hbase-snapshot/.tmp/snaptb0-testConsecutiveExports 2025-01-08T10:46:44,205 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-6488860248454022559.jar 2025-01-08T10:46:44,206 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:44,206 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:44,206 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:45,429 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-15496653336553088552.jar 2025-01-08T10:46:45,429 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:45,430 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:45,528 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-6506238688827376371.jar 2025-01-08T10:46:45,528 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:45,529 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:45,529 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:45,529 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:45,530 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:45,530 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T10:46:45,531 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T10:46:45,531 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T10:46:45,531 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T10:46:45,532 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T10:46:45,532 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T10:46:45,532 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T10:46:45,533 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T10:46:45,533 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T10:46:45,533 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T10:46:45,534 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T10:46:45,534 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T10:46:45,534 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T10:46:45,535 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:46:45,535 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:46:45,536 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:46:45,536 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:46:45,536 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:46:45,537 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:46:45,537 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:46:45,634 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742088_1264 (size=127628) 2025-01-08T10:46:45,635 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742088_1264 (size=127628) 2025-01-08T10:46:45,636 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742088_1264 (size=127628) 2025-01-08T10:46:45,684 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742089_1265 (size=2172137) 2025-01-08T10:46:45,684 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742089_1265 (size=2172137) 2025-01-08T10:46:45,685 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742089_1265 (size=2172137) 2025-01-08T10:46:45,732 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742090_1266 (size=213228) 2025-01-08T10:46:45,732 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742090_1266 (size=213228) 2025-01-08T10:46:45,733 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742090_1266 (size=213228) 2025-01-08T10:46:45,751 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742091_1267 (size=1877034) 2025-01-08T10:46:45,751 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742091_1267 (size=1877034) 2025-01-08T10:46:45,751 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742091_1267 (size=1877034) 2025-01-08T10:46:45,768 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742092_1268 (size=451756) 2025-01-08T10:46:45,768 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742092_1268 (size=451756) 2025-01-08T10:46:45,769 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742092_1268 (size=451756) 2025-01-08T10:46:45,778 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742093_1269 (size=533455) 2025-01-08T10:46:45,778 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742093_1269 (size=533455) 2025-01-08T10:46:45,779 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742093_1269 (size=533455) 2025-01-08T10:46:45,812 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742094_1270 (size=7280644) 2025-01-08T10:46:45,812 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742094_1270 (size=7280644) 2025-01-08T10:46:45,812 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742094_1270 (size=7280644) 2025-01-08T10:46:45,844 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742095_1271 (size=4188619) 2025-01-08T10:46:45,844 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742095_1271 (size=4188619) 2025-01-08T10:46:45,844 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742095_1271 (size=4188619) 2025-01-08T10:46:45,853 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742096_1272 (size=20406) 2025-01-08T10:46:45,854 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742096_1272 (size=20406) 2025-01-08T10:46:45,854 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742096_1272 (size=20406) 2025-01-08T10:46:45,863 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742097_1273 (size=75495) 2025-01-08T10:46:45,864 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742097_1273 (size=75495) 2025-01-08T10:46:45,865 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742097_1273 (size=75495) 2025-01-08T10:46:45,873 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742098_1274 (size=45609) 2025-01-08T10:46:45,873 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742098_1274 (size=45609) 2025-01-08T10:46:45,874 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742098_1274 (size=45609) 2025-01-08T10:46:45,889 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742099_1275 (size=110084) 2025-01-08T10:46:45,890 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742099_1275 (size=110084) 2025-01-08T10:46:45,890 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742099_1275 (size=110084) 2025-01-08T10:46:45,907 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742100_1276 (size=1323991) 2025-01-08T10:46:45,908 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742100_1276 (size=1323991) 2025-01-08T10:46:45,908 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742100_1276 (size=1323991) 2025-01-08T10:46:45,937 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742101_1277 (size=23076) 2025-01-08T10:46:45,938 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742101_1277 (size=23076) 2025-01-08T10:46:45,938 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742101_1277 (size=23076) 2025-01-08T10:46:45,959 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742102_1278 (size=126803) 2025-01-08T10:46:45,962 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742102_1278 (size=126803) 2025-01-08T10:46:45,963 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742102_1278 (size=126803) 2025-01-08T10:46:45,992 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742103_1279 (size=322274) 2025-01-08T10:46:45,993 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742103_1279 (size=322274) 2025-01-08T10:46:45,993 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742103_1279 (size=322274) 2025-01-08T10:46:46,039 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742104_1280 (size=6350714) 2025-01-08T10:46:46,040 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742104_1280 (size=6350714) 2025-01-08T10:46:46,040 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742104_1280 (size=6350714) 2025-01-08T10:46:46,065 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742105_1281 (size=912094) 2025-01-08T10:46:46,065 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742105_1281 (size=912094) 2025-01-08T10:46:46,065 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742105_1281 (size=912094) 2025-01-08T10:46:46,118 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742106_1282 (size=1832290) 2025-01-08T10:46:46,118 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742106_1282 (size=1832290) 2025-01-08T10:46:46,119 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742106_1282 (size=1832290) 2025-01-08T10:46:46,137 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742107_1283 (size=30081) 2025-01-08T10:46:46,137 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742107_1283 (size=30081) 2025-01-08T10:46:46,138 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742107_1283 (size=30081) 2025-01-08T10:46:46,159 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742108_1284 (size=53616) 2025-01-08T10:46:46,160 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742108_1284 (size=53616) 2025-01-08T10:46:46,160 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742108_1284 (size=53616) 2025-01-08T10:46:46,181 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742109_1285 (size=29229) 2025-01-08T10:46:46,181 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742109_1285 (size=29229) 2025-01-08T10:46:46,182 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742109_1285 (size=29229) 2025-01-08T10:46:46,211 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742110_1286 (size=169089) 2025-01-08T10:46:46,211 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742110_1286 (size=169089) 2025-01-08T10:46:46,211 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742110_1286 (size=169089) 2025-01-08T10:46:46,259 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742111_1287 (size=5175431) 2025-01-08T10:46:46,259 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742111_1287 (size=5175431) 2025-01-08T10:46:46,260 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742111_1287 (size=5175431) 2025-01-08T10:46:46,285 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742112_1288 (size=136454) 2025-01-08T10:46:46,285 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742112_1288 (size=136454) 2025-01-08T10:46:46,286 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742112_1288 (size=136454) 2025-01-08T10:46:46,317 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742113_1289 (size=3317408) 2025-01-08T10:46:46,318 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742113_1289 (size=3317408) 2025-01-08T10:46:46,318 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742113_1289 (size=3317408) 2025-01-08T10:46:46,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742114_1290 (size=503880) 2025-01-08T10:46:46,333 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742114_1290 (size=503880) 2025-01-08T10:46:46,333 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742114_1290 (size=503880) 2025-01-08T10:46:46,367 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742115_1291 (size=4695811) 2025-01-08T10:46:46,367 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742115_1291 (size=4695811) 2025-01-08T10:46:46,368 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742115_1291 (size=4695811) 2025-01-08T10:46:46,372 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T10:46:46,375 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'snaptb0-testConsecutiveExports' hfile list 2025-01-08T10:46:46,378 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(751): export split=0 size=33.2 K 2025-01-08T10:46:46,408 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742116_1292 (size=714) 2025-01-08T10:46:46,408 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742116_1292 (size=714) 2025-01-08T10:46:46,409 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742116_1292 (size=714) 2025-01-08T10:46:46,428 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742117_1293 (size=15) 2025-01-08T10:46:46,429 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742117_1293 (size=15) 2025-01-08T10:46:46,430 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742117_1293 (size=15) 2025-01-08T10:46:46,469 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742118_1294 (size=304976) 2025-01-08T10:46:46,469 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742118_1294 (size=304976) 2025-01-08T10:46:46,469 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742118_1294 (size=304976) 2025-01-08T10:46:47,184 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_1/usercache/jenkins/appcache/application_1736333116683_0004/container_1736333116683_0004_01_000002/launch_container.sh] 2025-01-08T10:46:47,184 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_1/usercache/jenkins/appcache/application_1736333116683_0004/container_1736333116683_0004_01_000002/container_tokens] 2025-01-08T10:46:47,184 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_1/usercache/jenkins/appcache/application_1736333116683_0004/container_1736333116683_0004_01_000002/sysfs] 2025-01-08T10:46:48,186 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0004_000001 (auth:SIMPLE) from 127.0.0.1:35082 2025-01-08T10:46:48,186 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T10:46:48,187 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T10:46:48,200 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_2/usercache/jenkins/appcache/application_1736333116683_0004/container_1736333116683_0004_01_000001/launch_container.sh] 2025-01-08T10:46:48,200 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_2/usercache/jenkins/appcache/application_1736333116683_0004/container_1736333116683_0004_01_000001/container_tokens] 2025-01-08T10:46:48,201 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_2/usercache/jenkins/appcache/application_1736333116683_0004/container_1736333116683_0004_01_000001/sysfs] 2025-01-08T10:46:48,892 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0005_000001 (auth:SIMPLE) from 127.0.0.1:57054 2025-01-08T10:46:55,976 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0005_000001 (auth:SIMPLE) from 127.0.0.1:40510 2025-01-08T10:46:56,334 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742119_1295 (size=350650) 2025-01-08T10:46:56,335 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742119_1295 (size=350650) 2025-01-08T10:46:56,336 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742119_1295 (size=350650) 2025-01-08T10:46:58,220 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0005_000001 (auth:SIMPLE) from 127.0.0.1:50986 2025-01-08T10:47:02,891 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742120_1296 (size=16925) 2025-01-08T10:47:02,892 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742120_1296 (size=16925) 2025-01-08T10:47:02,892 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742120_1296 (size=16925) 2025-01-08T10:47:02,919 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742121_1297 (size=462) 2025-01-08T10:47:02,919 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742121_1297 (size=462) 2025-01-08T10:47:02,920 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742121_1297 (size=462) 2025-01-08T10:47:02,925 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_2/usercache/jenkins/appcache/application_1736333116683_0005/container_1736333116683_0005_01_000002/launch_container.sh] 2025-01-08T10:47:02,925 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_2/usercache/jenkins/appcache/application_1736333116683_0005/container_1736333116683_0005_01_000002/container_tokens] 2025-01-08T10:47:02,925 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_2/usercache/jenkins/appcache/application_1736333116683_0005/container_1736333116683_0005_01_000002/sysfs] 2025-01-08T10:47:02,975 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742122_1298 (size=16925) 2025-01-08T10:47:02,975 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742122_1298 (size=16925) 2025-01-08T10:47:02,976 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742122_1298 (size=16925) 2025-01-08T10:47:03,034 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742123_1299 (size=350650) 2025-01-08T10:47:03,034 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742123_1299 (size=350650) 2025-01-08T10:47:03,034 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742123_1299 (size=350650) 2025-01-08T10:47:03,048 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0005_000001 (auth:SIMPLE) from 127.0.0.1:50994 2025-01-08T10:47:04,722 INFO [Time-limited test {}] snapshot.ExportSnapshot(1207): Finalize the Snapshot Export 2025-01-08T10:47:04,722 INFO [Time-limited test {}] snapshot.ExportSnapshot(1218): Verify the exported snapshot's expiration status and integrity. 2025-01-08T10:47:04,726 INFO [Time-limited test {}] snapshot.ExportSnapshot(1224): Export Completed: snaptb0-testConsecutiveExports 2025-01-08T10:47:04,726 INFO [Time-limited test {}] snapshot.TestExportSnapshot(362): Exported snapshot 2025-01-08T10:47:04,726 INFO [Time-limited test {}] snapshot.TestExportSnapshot(373): Verified filesystem state 2025-01-08T10:47:04,726 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-832679797_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testConsecutiveExports at hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testConsecutiveExports 2025-01-08T10:47:04,728 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testConsecutiveExports/.snapshotinfo 2025-01-08T10:47:04,728 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testConsecutiveExports/data.manifest 2025-01-08T10:47:04,728 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in org.apache.hadoop.fs.LocalFileSystem@5d4621f9 in root file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333186649/.hbase-snapshot/snaptb0-testConsecutiveExports at file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333186649/.hbase-snapshot/snaptb0-testConsecutiveExports 2025-01-08T10:47:04,728 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333186649/.hbase-snapshot/snaptb0-testConsecutiveExports/data.manifest 2025-01-08T10:47:04,728 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333186649/.hbase-snapshot/snaptb0-testConsecutiveExports/.snapshotinfo 2025-01-08T10:47:04,752 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testConsecutiveExports 2025-01-08T10:47:04,752 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.2 disable testtb-testConsecutiveExports 2025-01-08T10:47:04,753 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=92, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testConsecutiveExports 2025-01-08T10:47:04,755 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=92 2025-01-08T10:47:04,756 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testConsecutiveExports","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333224755"}]},"ts":"1736333224755"} 2025-01-08T10:47:04,757 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testConsecutiveExports, state=DISABLING in hbase:meta 2025-01-08T10:47:04,763 INFO [PEWorker-4 {}] procedure.DisableTableProcedure(284): Set testtb-testConsecutiveExports to state=DISABLING 2025-01-08T10:47:04,764 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=93, ppid=92, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testConsecutiveExports}] 2025-01-08T10:47:04,765 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=94, ppid=93, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=df917187823483f4cb65dabe50a92ccc, UNASSIGN}, {pid=95, ppid=93, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=677a448659d57404fc829eb39d04dfa9, UNASSIGN}] 2025-01-08T10:47:04,766 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=95, ppid=93, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=677a448659d57404fc829eb39d04dfa9, UNASSIGN 2025-01-08T10:47:04,766 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=94, ppid=93, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=df917187823483f4cb65dabe50a92ccc, UNASSIGN 2025-01-08T10:47:04,768 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=94 updating hbase:meta row=df917187823483f4cb65dabe50a92ccc, regionState=CLOSING, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:04,768 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=95 updating hbase:meta row=677a448659d57404fc829eb39d04dfa9, regionState=CLOSING, regionLocation=67f42ee69b4e,45003,1736333108039 2025-01-08T10:47:04,769 DEBUG [PEWorker-3 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T10:47:04,769 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=96, ppid=95, state=RUNNABLE; CloseRegionProcedure 677a448659d57404fc829eb39d04dfa9, server=67f42ee69b4e,45003,1736333108039}] 2025-01-08T10:47:04,770 DEBUG [PEWorker-2 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T10:47:04,770 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=97, ppid=94, state=RUNNABLE; CloseRegionProcedure df917187823483f4cb65dabe50a92ccc, server=67f42ee69b4e,36283,1736333108166}] 2025-01-08T10:47:04,856 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=92 2025-01-08T10:47:04,921 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:47:04,922 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] handler.UnassignRegionHandler(124): Close 677a448659d57404fc829eb39d04dfa9 2025-01-08T10:47:04,922 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T10:47:04,922 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:04,922 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] regionserver.HRegion(1681): Closing 677a448659d57404fc829eb39d04dfa9, disabling compactions & flushes 2025-01-08T10:47:04,922 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] regionserver.HRegion(1703): Closing region testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9. 2025-01-08T10:47:04,922 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9. 2025-01-08T10:47:04,922 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] regionserver.HRegion(1791): Acquired close lock on testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9. after waiting 0 ms 2025-01-08T10:47:04,922 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] regionserver.HRegion(1801): Updates disabled for region testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9. 2025-01-08T10:47:04,923 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] handler.UnassignRegionHandler(124): Close df917187823483f4cb65dabe50a92ccc 2025-01-08T10:47:04,923 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T10:47:04,923 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] regionserver.HRegion(1681): Closing df917187823483f4cb65dabe50a92ccc, disabling compactions & flushes 2025-01-08T10:47:04,923 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] regionserver.HRegion(1703): Closing region testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc. 2025-01-08T10:47:04,923 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc. 2025-01-08T10:47:04,923 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] regionserver.HRegion(1791): Acquired close lock on testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc. after waiting 0 ms 2025-01-08T10:47:04,923 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] regionserver.HRegion(1801): Updates disabled for region testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc. 2025-01-08T10:47:04,929 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/df917187823483f4cb65dabe50a92ccc/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T10:47:04,929 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/677a448659d57404fc829eb39d04dfa9/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T10:47:04,930 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:47:04,930 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] regionserver.HRegion(1922): Closed testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc. 2025-01-08T10:47:04,930 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] regionserver.HRegion(1635): Region close journal for df917187823483f4cb65dabe50a92ccc: 2025-01-08T10:47:04,930 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:47:04,930 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] regionserver.HRegion(1922): Closed testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9. 2025-01-08T10:47:04,930 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] regionserver.HRegion(1635): Region close journal for 677a448659d57404fc829eb39d04dfa9: 2025-01-08T10:47:04,931 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] handler.UnassignRegionHandler(170): Closed df917187823483f4cb65dabe50a92ccc 2025-01-08T10:47:04,932 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=94 updating hbase:meta row=df917187823483f4cb65dabe50a92ccc, regionState=CLOSED 2025-01-08T10:47:04,932 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] handler.UnassignRegionHandler(170): Closed 677a448659d57404fc829eb39d04dfa9 2025-01-08T10:47:04,933 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=95 updating hbase:meta row=677a448659d57404fc829eb39d04dfa9, regionState=CLOSED 2025-01-08T10:47:04,936 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=97, resume processing ppid=94 2025-01-08T10:47:04,936 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=97, ppid=94, state=SUCCESS; CloseRegionProcedure df917187823483f4cb65dabe50a92ccc, server=67f42ee69b4e,36283,1736333108166 in 163 msec 2025-01-08T10:47:04,936 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=96, resume processing ppid=95 2025-01-08T10:47:04,937 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=94, ppid=93, state=SUCCESS; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=df917187823483f4cb65dabe50a92ccc, UNASSIGN in 171 msec 2025-01-08T10:47:04,937 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=96, ppid=95, state=SUCCESS; CloseRegionProcedure 677a448659d57404fc829eb39d04dfa9, server=67f42ee69b4e,45003,1736333108039 in 165 msec 2025-01-08T10:47:04,938 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=95, resume processing ppid=93 2025-01-08T10:47:04,938 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=95, ppid=93, state=SUCCESS; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=677a448659d57404fc829eb39d04dfa9, UNASSIGN in 171 msec 2025-01-08T10:47:04,940 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=93, resume processing ppid=92 2025-01-08T10:47:04,940 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=93, ppid=92, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testConsecutiveExports in 174 msec 2025-01-08T10:47:04,941 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testConsecutiveExports","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333224941"}]},"ts":"1736333224941"} 2025-01-08T10:47:04,942 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testConsecutiveExports, state=DISABLED in hbase:meta 2025-01-08T10:47:04,954 INFO [PEWorker-5 {}] procedure.DisableTableProcedure(296): Set testtb-testConsecutiveExports to state=DISABLED 2025-01-08T10:47:04,956 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=92, state=SUCCESS; DisableTableProcedure table=testtb-testConsecutiveExports in 202 msec 2025-01-08T10:47:05,057 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=92 2025-01-08T10:47:05,058 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testConsecutiveExports, procId: 92 completed 2025-01-08T10:47:05,058 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.2 delete testtb-testConsecutiveExports 2025-01-08T10:47:05,059 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=98, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testConsecutiveExports 2025-01-08T10:47:05,060 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=98, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testConsecutiveExports 2025-01-08T10:47:05,060 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testConsecutiveExports 2025-01-08T10:47:05,061 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=98, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testConsecutiveExports 2025-01-08T10:47:05,062 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36283 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testConsecutiveExports 2025-01-08T10:47:05,064 DEBUG [HFileArchiver-9 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/df917187823483f4cb65dabe50a92ccc 2025-01-08T10:47:05,066 DEBUG [HFileArchiver-10 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/677a448659d57404fc829eb39d04dfa9 2025-01-08T10:47:05,067 DEBUG [HFileArchiver-9 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/df917187823483f4cb65dabe50a92ccc/cf, FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/df917187823483f4cb65dabe50a92ccc/recovered.edits] 2025-01-08T10:47:05,069 DEBUG [HFileArchiver-10 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/677a448659d57404fc829eb39d04dfa9/cf, FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/677a448659d57404fc829eb39d04dfa9/recovered.edits] 2025-01-08T10:47:05,072 DEBUG [HFileArchiver-9 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/df917187823483f4cb65dabe50a92ccc/cf/6112f6e037fd410c84ce02256175994f to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testConsecutiveExports/df917187823483f4cb65dabe50a92ccc/cf/6112f6e037fd410c84ce02256175994f 2025-01-08T10:47:05,073 DEBUG [HFileArchiver-10 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/677a448659d57404fc829eb39d04dfa9/cf/fc8a56c09f1a486fbad516fee9a400d4 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testConsecutiveExports/677a448659d57404fc829eb39d04dfa9/cf/fc8a56c09f1a486fbad516fee9a400d4 2025-01-08T10:47:05,075 DEBUG [HFileArchiver-9 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/df917187823483f4cb65dabe50a92ccc/recovered.edits/9.seqid to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testConsecutiveExports/df917187823483f4cb65dabe50a92ccc/recovered.edits/9.seqid 2025-01-08T10:47:05,075 DEBUG [HFileArchiver-9 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/df917187823483f4cb65dabe50a92ccc 2025-01-08T10:47:05,076 DEBUG [HFileArchiver-10 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/677a448659d57404fc829eb39d04dfa9/recovered.edits/9.seqid to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testConsecutiveExports/677a448659d57404fc829eb39d04dfa9/recovered.edits/9.seqid 2025-01-08T10:47:05,077 DEBUG [HFileArchiver-10 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testConsecutiveExports/677a448659d57404fc829eb39d04dfa9 2025-01-08T10:47:05,077 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(313): Archived testtb-testConsecutiveExports regions 2025-01-08T10:47:05,077 DEBUG [PEWorker-2 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac 2025-01-08T10:47:05,078 DEBUG [PEWorker-2 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf] 2025-01-08T10:47:05,082 DEBUG [PEWorker-2 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/c4ca4238a0b923820dcc509a6f75849b202501087c52c1f04b5d4722a0c956a8eccc0d94_677a448659d57404fc829eb39d04dfa9 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/c4ca4238a0b923820dcc509a6f75849b202501087c52c1f04b5d4722a0c956a8eccc0d94_677a448659d57404fc829eb39d04dfa9 2025-01-08T10:47:05,083 DEBUG [PEWorker-2 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/d41d8cd98f00b204e9800998ecf8427e20250108037a5f0bd72a48acb4762718720e5abf_df917187823483f4cb65dabe50a92ccc to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/d41d8cd98f00b204e9800998ecf8427e20250108037a5f0bd72a48acb4762718720e5abf_df917187823483f4cb65dabe50a92ccc 2025-01-08T10:47:05,083 DEBUG [PEWorker-2 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac 2025-01-08T10:47:05,085 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=98, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testConsecutiveExports 2025-01-08T10:47:05,088 WARN [PEWorker-2 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testtb-testConsecutiveExports from hbase:meta 2025-01-08T10:47:05,100 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testConsecutiveExports 2025-01-08T10:47:05,100 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testConsecutiveExports 2025-01-08T10:47:05,100 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testConsecutiveExports 2025-01-08T10:47:05,100 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testConsecutiveExports 2025-01-08T10:47:05,101 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testConsecutiveExports with data PBUF 2025-01-08T10:47:05,101 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testConsecutiveExports with data PBUF 2025-01-08T10:47:05,101 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testConsecutiveExports with data PBUF 2025-01-08T10:47:05,101 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testConsecutiveExports with data PBUF 2025-01-08T10:47:05,102 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testConsecutiveExports' descriptor. 2025-01-08T10:47:05,103 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=98, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testConsecutiveExports 2025-01-08T10:47:05,103 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testConsecutiveExports' from region states. 2025-01-08T10:47:05,103 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736333225103"}]},"ts":"9223372036854775807"} 2025-01-08T10:47:05,103 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736333225103"}]},"ts":"9223372036854775807"} 2025-01-08T10:47:05,105 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T10:47:05,105 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => df917187823483f4cb65dabe50a92ccc, NAME => 'testtb-testConsecutiveExports,,1736333184157.df917187823483f4cb65dabe50a92ccc.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => 677a448659d57404fc829eb39d04dfa9, NAME => 'testtb-testConsecutiveExports,1,1736333184157.677a448659d57404fc829eb39d04dfa9.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T10:47:05,105 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testConsecutiveExports' as deleted. 2025-01-08T10:47:05,105 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testConsecutiveExports","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736333225105"}]},"ts":"9223372036854775807"} 2025-01-08T10:47:05,107 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testConsecutiveExports state from META 2025-01-08T10:47:05,138 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testConsecutiveExports 2025-01-08T10:47:05,138 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testConsecutiveExports 2025-01-08T10:47:05,138 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testConsecutiveExports 2025-01-08T10:47:05,138 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:05,138 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:05,138 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:05,138 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testConsecutiveExports 2025-01-08T10:47:05,138 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:05,139 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=98 2025-01-08T10:47:05,162 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(133): Finished pid=98, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testConsecutiveExports 2025-01-08T10:47:05,163 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=98, state=SUCCESS; DeleteTableProcedure table=testtb-testConsecutiveExports in 104 msec 2025-01-08T10:47:05,240 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=98 2025-01-08T10:47:05,241 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testConsecutiveExports, procId: 98 completed 2025-01-08T10:47:05,248 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.2 delete name: "emptySnaptb0-testConsecutiveExports" 2025-01-08T10:47:05,250 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(380): Deleting snapshot: emptySnaptb0-testConsecutiveExports 2025-01-08T10:47:05,251 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.2 delete name: "snaptb0-testConsecutiveExports" 2025-01-08T10:47:05,253 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testConsecutiveExports 2025-01-08T10:47:05,278 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: snapshot.TestMobSecureExportSnapshot#testConsecutiveExports Thread=787 (was 785) Potentially hanging thread: ApplicationMasterLauncher #5 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-9 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #7 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_1051535197_1 at /127.0.0.1:49798 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (175901922) connection to localhost/127.0.0.1:37695 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: ApplicationMasterLauncher #8 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:37695 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-25 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-10 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_1051535197_1 at /127.0.0.1:39728 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-832679797_22 at /127.0.0.1:50842 [Waiting for operation #3] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #6 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-23 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-24 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LogDeleter #2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1177) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-22 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: process reaper (pid 17347) java.base@17.0.11/java.lang.ProcessHandleImpl.waitForProcessExit0(Native Method) java.base@17.0.11/java.lang.ProcessHandleImpl$1.run(ProcessHandleImpl.java:150) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-832679797_22 at /127.0.0.1:59176 [Waiting for operation #3] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Thread-4079 java.base@17.0.11/java.io.FileInputStream.readBytes(Native Method) java.base@17.0.11/java.io.FileInputStream.read(FileInputStream.java:276) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:282) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) java.base@17.0.11/sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:281) java.base@17.0.11/sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:324) java.base@17.0.11/sun.nio.cs.StreamDecoder.read(StreamDecoder.java:189) java.base@17.0.11/java.io.InputStreamReader.read(InputStreamReader.java:177) java.base@17.0.11/java.io.BufferedReader.fill(BufferedReader.java:162) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:329) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:396) app//org.apache.hadoop.util.Shell$1.run(Shell.java:1025) - Thread LEAK? -, OpenFileDescriptor=803 (was 803), MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=841 (was 657) - SystemLoadAverage LEAK? -, ProcessCount=18 (was 18), AvailableMemoryMB=1473 (was 2025) 2025-01-08T10:47:05,278 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=787 is superior to 500 2025-01-08T10:47:05,299 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: snapshot.TestMobSecureExportSnapshot#testExportFileSystemStateWithMergeRegion Thread=787, OpenFileDescriptor=803, MaxFileDescriptor=1048576, SystemLoadAverage=841, ProcessCount=18, AvailableMemoryMB=1473 2025-01-08T10:47:05,299 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=787 is superior to 500 2025-01-08T10:47:05,301 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.2 create 'testtb-testExportFileSystemStateWithMergeRegion', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T10:47:05,302 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=99, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:05,303 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=99, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T10:47:05,303 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "testtb-testExportFileSystemStateWithMergeRegion" procId is: 99 2025-01-08T10:47:05,304 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=99, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T10:47:05,304 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=99 2025-01-08T10:47:05,311 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742124_1300 (size=458) 2025-01-08T10:47:05,311 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742124_1300 (size=458) 2025-01-08T10:47:05,312 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742124_1300 (size=458) 2025-01-08T10:47:05,314 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 4150f0b85bebbb7094b7a1d91b0a6f32, NAME => 'testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testtb-testExportFileSystemStateWithMergeRegion', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:47:05,314 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => 6ff7afa9ac49884297a386018fda40ef, NAME => 'testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testtb-testExportFileSystemStateWithMergeRegion', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:47:05,327 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742125_1301 (size=83) 2025-01-08T10:47:05,328 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742125_1301 (size=83) 2025-01-08T10:47:05,328 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742125_1301 (size=83) 2025-01-08T10:47:05,328 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:47:05,329 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-1 {}] regionserver.HRegion(1681): Closing 6ff7afa9ac49884297a386018fda40ef, disabling compactions & flushes 2025-01-08T10:47:05,329 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef. 2025-01-08T10:47:05,329 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef. 2025-01-08T10:47:05,329 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef. after waiting 0 ms 2025-01-08T10:47:05,329 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef. 2025-01-08T10:47:05,329 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef. 2025-01-08T10:47:05,329 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-1 {}] regionserver.HRegion(1635): Region close journal for 6ff7afa9ac49884297a386018fda40ef: 2025-01-08T10:47:05,335 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742126_1302 (size=83) 2025-01-08T10:47:05,336 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742126_1302 (size=83) 2025-01-08T10:47:05,336 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742126_1302 (size=83) 2025-01-08T10:47:05,405 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=99 2025-01-08T10:47:05,607 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=99 2025-01-08T10:47:05,714 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T10:47:05,736 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:47:05,736 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-0 {}] regionserver.HRegion(1681): Closing 4150f0b85bebbb7094b7a1d91b0a6f32, disabling compactions & flushes 2025-01-08T10:47:05,736 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32. 2025-01-08T10:47:05,736 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32. 2025-01-08T10:47:05,736 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32. after waiting 0 ms 2025-01-08T10:47:05,736 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32. 2025-01-08T10:47:05,736 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32. 2025-01-08T10:47:05,736 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-0 {}] regionserver.HRegion(1635): Region close journal for 4150f0b85bebbb7094b7a1d91b0a6f32: 2025-01-08T10:47:05,740 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=99, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T10:47:05,740 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef.","families":{"info":[{"qualifier":"regioninfo","vlen":82,"tag":[],"timestamp":"1736333225740"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736333225740"}]},"ts":"1736333225740"} 2025-01-08T10:47:05,740 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32.","families":{"info":[{"qualifier":"regioninfo","vlen":82,"tag":[],"timestamp":"1736333225740"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736333225740"}]},"ts":"1736333225740"} 2025-01-08T10:47:05,745 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T10:47:05,746 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=99, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T10:47:05,746 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333225746"}]},"ts":"1736333225746"} 2025-01-08T10:47:05,747 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithMergeRegion, state=ENABLING in hbase:meta 2025-01-08T10:47:05,769 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(202): Hosts are {67f42ee69b4e=0} racks are {/default-rack=0} 2025-01-08T10:47:05,771 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T10:47:05,771 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T10:47:05,771 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T10:47:05,771 INFO [PEWorker-3 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T10:47:05,771 INFO [PEWorker-3 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T10:47:05,771 INFO [PEWorker-3 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T10:47:05,771 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T10:47:05,771 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=100, ppid=99, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=4150f0b85bebbb7094b7a1d91b0a6f32, ASSIGN}, {pid=101, ppid=99, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=6ff7afa9ac49884297a386018fda40ef, ASSIGN}] 2025-01-08T10:47:05,772 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=101, ppid=99, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=6ff7afa9ac49884297a386018fda40ef, ASSIGN 2025-01-08T10:47:05,772 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=100, ppid=99, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=4150f0b85bebbb7094b7a1d91b0a6f32, ASSIGN 2025-01-08T10:47:05,773 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=100, ppid=99, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=4150f0b85bebbb7094b7a1d91b0a6f32, ASSIGN; state=OFFLINE, location=67f42ee69b4e,38323,1736333108259; forceNewPlan=false, retain=false 2025-01-08T10:47:05,773 INFO [PEWorker-1 {}] assignment.TransitRegionStateProcedure(264): Starting pid=101, ppid=99, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=6ff7afa9ac49884297a386018fda40ef, ASSIGN; state=OFFLINE, location=67f42ee69b4e,45003,1736333108039; forceNewPlan=false, retain=false 2025-01-08T10:47:05,908 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=99 2025-01-08T10:47:05,924 INFO [67f42ee69b4e:34551 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T10:47:05,924 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=101 updating hbase:meta row=6ff7afa9ac49884297a386018fda40ef, regionState=OPENING, regionLocation=67f42ee69b4e,45003,1736333108039 2025-01-08T10:47:05,924 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=100 updating hbase:meta row=4150f0b85bebbb7094b7a1d91b0a6f32, regionState=OPENING, regionLocation=67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:05,926 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=102, ppid=100, state=RUNNABLE; OpenRegionProcedure 4150f0b85bebbb7094b7a1d91b0a6f32, server=67f42ee69b4e,38323,1736333108259}] 2025-01-08T10:47:05,927 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=103, ppid=101, state=RUNNABLE; OpenRegionProcedure 6ff7afa9ac49884297a386018fda40ef, server=67f42ee69b4e,45003,1736333108039}] 2025-01-08T10:47:06,078 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:06,079 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:47:06,082 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] handler.AssignRegionHandler(135): Open testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef. 2025-01-08T10:47:06,082 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] handler.AssignRegionHandler(135): Open testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32. 2025-01-08T10:47:06,082 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(7285): Opening region: {ENCODED => 6ff7afa9ac49884297a386018fda40ef, NAME => 'testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T10:47:06,083 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(7285): Opening region: {ENCODED => 4150f0b85bebbb7094b7a1d91b0a6f32, NAME => 'testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T10:47:06,083 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef. service=AccessControlService 2025-01-08T10:47:06,083 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32. service=AccessControlService 2025-01-08T10:47:06,083 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:47:06,083 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:47:06,083 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportFileSystemStateWithMergeRegion 6ff7afa9ac49884297a386018fda40ef 2025-01-08T10:47:06,083 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportFileSystemStateWithMergeRegion 4150f0b85bebbb7094b7a1d91b0a6f32 2025-01-08T10:47:06,083 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:47:06,083 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:47:06,083 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(7327): checking encryption for 6ff7afa9ac49884297a386018fda40ef 2025-01-08T10:47:06,083 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(7327): checking encryption for 4150f0b85bebbb7094b7a1d91b0a6f32 2025-01-08T10:47:06,083 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(7330): checking classloading for 6ff7afa9ac49884297a386018fda40ef 2025-01-08T10:47:06,083 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(7330): checking classloading for 4150f0b85bebbb7094b7a1d91b0a6f32 2025-01-08T10:47:06,085 INFO [StoreOpener-6ff7afa9ac49884297a386018fda40ef-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 6ff7afa9ac49884297a386018fda40ef 2025-01-08T10:47:06,085 INFO [StoreOpener-4150f0b85bebbb7094b7a1d91b0a6f32-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 4150f0b85bebbb7094b7a1d91b0a6f32 2025-01-08T10:47:06,087 INFO [StoreOpener-6ff7afa9ac49884297a386018fda40ef-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 6ff7afa9ac49884297a386018fda40ef columnFamilyName cf 2025-01-08T10:47:06,087 INFO [StoreOpener-4150f0b85bebbb7094b7a1d91b0a6f32-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 4150f0b85bebbb7094b7a1d91b0a6f32 columnFamilyName cf 2025-01-08T10:47:06,088 DEBUG [StoreOpener-6ff7afa9ac49884297a386018fda40ef-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:06,088 DEBUG [StoreOpener-4150f0b85bebbb7094b7a1d91b0a6f32-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:06,088 INFO [StoreOpener-6ff7afa9ac49884297a386018fda40ef-1 {}] regionserver.HStore(327): Store=6ff7afa9ac49884297a386018fda40ef/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:47:06,088 INFO [StoreOpener-4150f0b85bebbb7094b7a1d91b0a6f32-1 {}] regionserver.HStore(327): Store=4150f0b85bebbb7094b7a1d91b0a6f32/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:47:06,089 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/4150f0b85bebbb7094b7a1d91b0a6f32 2025-01-08T10:47:06,089 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/6ff7afa9ac49884297a386018fda40ef 2025-01-08T10:47:06,090 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/4150f0b85bebbb7094b7a1d91b0a6f32 2025-01-08T10:47:06,090 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/6ff7afa9ac49884297a386018fda40ef 2025-01-08T10:47:06,092 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(1085): writing seq id for 6ff7afa9ac49884297a386018fda40ef 2025-01-08T10:47:06,092 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(1085): writing seq id for 4150f0b85bebbb7094b7a1d91b0a6f32 2025-01-08T10:47:06,095 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/6ff7afa9ac49884297a386018fda40ef/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:47:06,095 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/4150f0b85bebbb7094b7a1d91b0a6f32/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:47:06,095 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(1102): Opened 6ff7afa9ac49884297a386018fda40ef; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=60541593, jitterRate=-0.09785996377468109}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T10:47:06,096 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(1001): Region open journal for 6ff7afa9ac49884297a386018fda40ef: 2025-01-08T10:47:06,097 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(1102): Opened 4150f0b85bebbb7094b7a1d91b0a6f32; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=67394042, jitterRate=0.004249483346939087}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T10:47:06,097 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(1001): Region open journal for 4150f0b85bebbb7094b7a1d91b0a6f32: 2025-01-08T10:47:06,097 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef., pid=103, masterSystemTime=1736333226079 2025-01-08T10:47:06,098 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32., pid=102, masterSystemTime=1736333226078 2025-01-08T10:47:06,098 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef. 2025-01-08T10:47:06,098 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] handler.AssignRegionHandler(164): Opened testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef. 2025-01-08T10:47:06,099 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=101 updating hbase:meta row=6ff7afa9ac49884297a386018fda40ef, regionState=OPEN, openSeqNum=2, regionLocation=67f42ee69b4e,45003,1736333108039 2025-01-08T10:47:06,099 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32. 2025-01-08T10:47:06,099 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] handler.AssignRegionHandler(164): Opened testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32. 2025-01-08T10:47:06,100 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=100 updating hbase:meta row=4150f0b85bebbb7094b7a1d91b0a6f32, regionState=OPEN, openSeqNum=2, regionLocation=67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:06,103 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=103, resume processing ppid=101 2025-01-08T10:47:06,104 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=103, ppid=101, state=SUCCESS; OpenRegionProcedure 6ff7afa9ac49884297a386018fda40ef, server=67f42ee69b4e,45003,1736333108039 in 174 msec 2025-01-08T10:47:06,104 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=102, resume processing ppid=100 2025-01-08T10:47:06,104 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=102, ppid=100, state=SUCCESS; OpenRegionProcedure 4150f0b85bebbb7094b7a1d91b0a6f32, server=67f42ee69b4e,38323,1736333108259 in 176 msec 2025-01-08T10:47:06,105 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=101, ppid=99, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=6ff7afa9ac49884297a386018fda40ef, ASSIGN in 333 msec 2025-01-08T10:47:06,106 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=100, resume processing ppid=99 2025-01-08T10:47:06,106 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=100, ppid=99, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=4150f0b85bebbb7094b7a1d91b0a6f32, ASSIGN in 333 msec 2025-01-08T10:47:06,107 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=99, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T10:47:06,107 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333226107"}]},"ts":"1736333226107"} 2025-01-08T10:47:06,108 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithMergeRegion, state=ENABLED in hbase:meta 2025-01-08T10:47:06,139 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=99, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T10:47:06,140 DEBUG [PEWorker-1 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testExportFileSystemStateWithMergeRegion jenkins: RWXCA 2025-01-08T10:47:06,142 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36283 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemStateWithMergeRegion], kv [jenkins: RWXCA] 2025-01-08T10:47:06,162 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:06,162 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:06,162 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:06,162 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:06,193 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:06,193 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:06,195 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=99, state=SUCCESS; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion in 892 msec 2025-01-08T10:47:06,203 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:06,203 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:06,409 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=99 2025-01-08T10:47:06,410 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion, procId: 99 completed 2025-01-08T10:47:06,413 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:06,413 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32. 2025-01-08T10:47:06,413 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T10:47:06,433 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.2 snapshot request for:{ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } 2025-01-08T10:47:06,433 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736333226433 (current time:1736333226433). 2025-01-08T10:47:06,433 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T10:47:06,433 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot emptySnaptb0-testExportFileSystemStateWithMergeRegion VERSION not specified, setting to 2 2025-01-08T10:47:06,433 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T10:47:06,434 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x0211e39f to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@4e630ee 2025-01-08T10:47:06,442 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@73be1f6e, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:47:06,443 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:47:06,444 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:46692, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:47:06,445 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x0211e39f to 127.0.0.1:52367 2025-01-08T10:47:06,446 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:47:06,447 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x4ceea5bf to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@79ba8d8c 2025-01-08T10:47:06,470 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7a7ca8ba, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:47:06,472 DEBUG [hconnection-0x4efe9af8-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:47:06,473 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:46694, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:47:06,477 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x4ceea5bf to 127.0.0.1:52367 2025-01-08T10:47:06,477 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:47:06,478 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemStateWithMergeRegion], kv [jenkins: RWXCA] 2025-01-08T10:47:06,478 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T10:47:06,479 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=104, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } 2025-01-08T10:47:06,480 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 }, snapshot procedure id = 104 2025-01-08T10:47:06,480 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T10:47:06,481 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=104 2025-01-08T10:47:06,481 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T10:47:06,484 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T10:47:06,493 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742127_1303 (size=215) 2025-01-08T10:47:06,493 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742127_1303 (size=215) 2025-01-08T10:47:06,493 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742127_1303 (size=215) 2025-01-08T10:47:06,494 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T10:47:06,494 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=105, ppid=104, state=RUNNABLE; SnapshotRegionProcedure 4150f0b85bebbb7094b7a1d91b0a6f32}, {pid=106, ppid=104, state=RUNNABLE; SnapshotRegionProcedure 6ff7afa9ac49884297a386018fda40ef}] 2025-01-08T10:47:06,495 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=106, ppid=104, state=RUNNABLE; SnapshotRegionProcedure 6ff7afa9ac49884297a386018fda40ef 2025-01-08T10:47:06,495 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=105, ppid=104, state=RUNNABLE; SnapshotRegionProcedure 4150f0b85bebbb7094b7a1d91b0a6f32 2025-01-08T10:47:06,582 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=104 2025-01-08T10:47:06,647 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:47:06,647 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:06,647 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=38323 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=105 2025-01-08T10:47:06,647 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=45003 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=106 2025-01-08T10:47:06,647 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=106}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef. 2025-01-08T10:47:06,648 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=106}] regionserver.HRegion(2538): Flush status journal for 6ff7afa9ac49884297a386018fda40ef: 2025-01-08T10:47:06,648 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=105}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32. 2025-01-08T10:47:06,648 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=106}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef. for emptySnaptb0-testExportFileSystemStateWithMergeRegion completed. 2025-01-08T10:47:06,648 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=105}] regionserver.HRegion(2538): Flush status journal for 4150f0b85bebbb7094b7a1d91b0a6f32: 2025-01-08T10:47:06,648 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=106}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef.' region-info for snapshot=emptySnaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:06,648 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=105}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32. for emptySnaptb0-testExportFileSystemStateWithMergeRegion completed. 2025-01-08T10:47:06,648 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=106}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:47:06,648 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=106}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T10:47:06,648 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=105}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32.' region-info for snapshot=emptySnaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:06,648 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=105}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:47:06,648 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=105}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T10:47:06,655 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742128_1304 (size=86) 2025-01-08T10:47:06,655 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742128_1304 (size=86) 2025-01-08T10:47:06,656 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742129_1305 (size=86) 2025-01-08T10:47:06,656 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742129_1305 (size=86) 2025-01-08T10:47:06,656 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742128_1304 (size=86) 2025-01-08T10:47:06,656 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742129_1305 (size=86) 2025-01-08T10:47:06,656 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=105}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32. 2025-01-08T10:47:06,656 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=105}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=105 2025-01-08T10:47:06,656 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=105 2025-01-08T10:47:06,657 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportFileSystemStateWithMergeRegion on region 4150f0b85bebbb7094b7a1d91b0a6f32 2025-01-08T10:47:06,657 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=105, ppid=104, state=RUNNABLE; SnapshotRegionProcedure 4150f0b85bebbb7094b7a1d91b0a6f32 2025-01-08T10:47:06,658 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=105, ppid=104, state=SUCCESS; SnapshotRegionProcedure 4150f0b85bebbb7094b7a1d91b0a6f32 in 163 msec 2025-01-08T10:47:06,783 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=104 2025-01-08T10:47:07,057 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=106}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef. 2025-01-08T10:47:07,057 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=106}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=106 2025-01-08T10:47:07,057 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=106 2025-01-08T10:47:07,057 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportFileSystemStateWithMergeRegion on region 6ff7afa9ac49884297a386018fda40ef 2025-01-08T10:47:07,058 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=106, ppid=104, state=RUNNABLE; SnapshotRegionProcedure 6ff7afa9ac49884297a386018fda40ef 2025-01-08T10:47:07,063 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=106, resume processing ppid=104 2025-01-08T10:47:07,063 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=106, ppid=104, state=SUCCESS; SnapshotRegionProcedure 6ff7afa9ac49884297a386018fda40ef in 567 msec 2025-01-08T10:47:07,063 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T10:47:07,063 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T10:47:07,064 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T10:47:07,064 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T10:47:07,064 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:07,065 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(328): No files under family: cf 2025-01-08T10:47:07,071 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742130_1306 (size=78) 2025-01-08T10:47:07,071 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742130_1306 (size=78) 2025-01-08T10:47:07,071 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742130_1306 (size=78) 2025-01-08T10:47:07,072 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T10:47:07,072 DEBUG [PEWorker-5 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for emptySnaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:07,073 DEBUG [PEWorker-5 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/emptySnaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:07,082 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742131_1307 (size=713) 2025-01-08T10:47:07,082 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742131_1307 (size=713) 2025-01-08T10:47:07,082 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742131_1307 (size=713) 2025-01-08T10:47:07,084 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=104 2025-01-08T10:47:07,088 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T10:47:07,092 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T10:47:07,093 DEBUG [PEWorker-5 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/emptySnaptb0-testExportFileSystemStateWithMergeRegion to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/emptySnaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:07,094 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T10:47:07,094 DEBUG [PEWorker-5 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 }, snapshot procedure id = 104 2025-01-08T10:47:07,095 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=104, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } in 615 msec 2025-01-08T10:47:07,585 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=104 2025-01-08T10:47:07,586 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion, procId: 104 completed 2025-01-08T10:47:07,589 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38323 {}] regionserver.HRegion(8254): writing data to region testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T10:47:07,589 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=45003 {}] regionserver.HRegion(8254): writing data to region testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T10:47:07,592 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:07,592 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32. 2025-01-08T10:47:07,593 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T10:47:07,595 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:07,596 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemStateWithMergeRegion Metrics about Tables on a single HBase RegionServer 2025-01-08T10:47:07,596 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testConsecutiveExports 2025-01-08T10:47:07,605 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.2 snapshot request for:{ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } 2025-01-08T10:47:07,605 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736333227605 (current time:1736333227605). 2025-01-08T10:47:07,605 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T10:47:07,605 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testExportFileSystemStateWithMergeRegion VERSION not specified, setting to 2 2025-01-08T10:47:07,605 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T10:47:07,606 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x455962c9 to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@7fca09c0 2025-01-08T10:47:07,609 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@28f749d9, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:47:07,611 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:47:07,612 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:46704, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:47:07,613 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x455962c9 to 127.0.0.1:52367 2025-01-08T10:47:07,613 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:47:07,614 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x05288d15 to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@2a830079 2025-01-08T10:47:07,616 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@38a83d71, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:47:07,618 DEBUG [hconnection-0x7cc93ece-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:47:07,619 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:46706, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:47:07,621 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x05288d15 to 127.0.0.1:52367 2025-01-08T10:47:07,621 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:47:07,622 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemStateWithMergeRegion], kv [jenkins: RWXCA] 2025-01-08T10:47:07,622 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T10:47:07,623 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=107, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } 2025-01-08T10:47:07,623 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 }, snapshot procedure id = 107 2025-01-08T10:47:07,624 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T10:47:07,624 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=107 2025-01-08T10:47:07,625 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T10:47:07,627 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T10:47:07,635 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742132_1308 (size=210) 2025-01-08T10:47:07,635 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742132_1308 (size=210) 2025-01-08T10:47:07,635 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742132_1308 (size=210) 2025-01-08T10:47:07,636 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T10:47:07,636 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=108, ppid=107, state=RUNNABLE; SnapshotRegionProcedure 4150f0b85bebbb7094b7a1d91b0a6f32}, {pid=109, ppid=107, state=RUNNABLE; SnapshotRegionProcedure 6ff7afa9ac49884297a386018fda40ef}] 2025-01-08T10:47:07,637 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=108, ppid=107, state=RUNNABLE; SnapshotRegionProcedure 4150f0b85bebbb7094b7a1d91b0a6f32 2025-01-08T10:47:07,637 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=109, ppid=107, state=RUNNABLE; SnapshotRegionProcedure 6ff7afa9ac49884297a386018fda40ef 2025-01-08T10:47:07,725 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=107 2025-01-08T10:47:07,788 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:07,788 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:47:07,789 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=45003 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=109 2025-01-08T10:47:07,789 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=38323 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=108 2025-01-08T10:47:07,789 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef. 2025-01-08T10:47:07,789 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32. 2025-01-08T10:47:07,789 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.HRegion(2837): Flushing 4150f0b85bebbb7094b7a1d91b0a6f32 1/1 column families, dataSize=199 B heapSize=688 B 2025-01-08T10:47:07,789 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.HRegion(2837): Flushing 6ff7afa9ac49884297a386018fda40ef 1/1 column families, dataSize=3.06 KB heapSize=6.86 KB 2025-01-08T10:47:07,808 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108df8d023a67d946d4a1716e125356d5e2_4150f0b85bebbb7094b7a1d91b0a6f32 is 71, key is 04fd8b33ae1f51945db3f9fdf0e66830/cf:q/1736333227589/Put/seqid=0 2025-01-08T10:47:07,810 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b2025010810301e755fe74f01a4220a74bb7a8296_6ff7afa9ac49884297a386018fda40ef is 71, key is 165da81f047150588a9e55721b93f357/cf:q/1736333227589/Put/seqid=0 2025-01-08T10:47:07,814 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742133_1309 (size=5101) 2025-01-08T10:47:07,814 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742133_1309 (size=5101) 2025-01-08T10:47:07,815 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742133_1309 (size=5101) 2025-01-08T10:47:07,815 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:07,821 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742134_1310 (size=8171) 2025-01-08T10:47:07,821 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742134_1310 (size=8171) 2025-01-08T10:47:07,821 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742134_1310 (size=8171) 2025-01-08T10:47:07,823 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:07,826 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108df8d023a67d946d4a1716e125356d5e2_4150f0b85bebbb7094b7a1d91b0a6f32 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/d41d8cd98f00b204e9800998ecf8427e20250108df8d023a67d946d4a1716e125356d5e2_4150f0b85bebbb7094b7a1d91b0a6f32 2025-01-08T10:47:07,827 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b2025010810301e755fe74f01a4220a74bb7a8296_6ff7afa9ac49884297a386018fda40ef to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/c4ca4238a0b923820dcc509a6f75849b2025010810301e755fe74f01a4220a74bb7a8296_6ff7afa9ac49884297a386018fda40ef 2025-01-08T10:47:07,827 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/4150f0b85bebbb7094b7a1d91b0a6f32/.tmp/cf/b85d87029a5646189c7a994b7c99cdba, store: [table=testtb-testExportFileSystemStateWithMergeRegion family=cf region=4150f0b85bebbb7094b7a1d91b0a6f32] 2025-01-08T10:47:07,827 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/6ff7afa9ac49884297a386018fda40ef/.tmp/cf/cd90586f90f64d5691980aaf0b4c9136, store: [table=testtb-testExportFileSystemStateWithMergeRegion family=cf region=6ff7afa9ac49884297a386018fda40ef] 2025-01-08T10:47:07,828 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/4150f0b85bebbb7094b7a1d91b0a6f32/.tmp/cf/b85d87029a5646189c7a994b7c99cdba is 224, key is 09c55efd1e6489d4c4b0f395c4a0b1cd5/cf:q/1736333227589/Put/seqid=0 2025-01-08T10:47:07,828 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/6ff7afa9ac49884297a386018fda40ef/.tmp/cf/cd90586f90f64d5691980aaf0b4c9136 is 224, key is 16b7764cd9e76a82be9575d278a35bf65/cf:q/1736333227589/Put/seqid=0 2025-01-08T10:47:07,837 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742136_1312 (size=15717) 2025-01-08T10:47:07,838 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742136_1312 (size=15717) 2025-01-08T10:47:07,838 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742136_1312 (size=15717) 2025-01-08T10:47:07,838 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742135_1311 (size=5976) 2025-01-08T10:47:07,838 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742135_1311 (size=5976) 2025-01-08T10:47:07,838 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742135_1311 (size=5976) 2025-01-08T10:47:07,838 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=3.1 K, hasBloomFilter=true, into tmp file hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/6ff7afa9ac49884297a386018fda40ef/.tmp/cf/cd90586f90f64d5691980aaf0b4c9136 2025-01-08T10:47:07,839 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=199, hasBloomFilter=true, into tmp file hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/4150f0b85bebbb7094b7a1d91b0a6f32/.tmp/cf/b85d87029a5646189c7a994b7c99cdba 2025-01-08T10:47:07,843 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/6ff7afa9ac49884297a386018fda40ef/.tmp/cf/cd90586f90f64d5691980aaf0b4c9136 as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/6ff7afa9ac49884297a386018fda40ef/cf/cd90586f90f64d5691980aaf0b4c9136 2025-01-08T10:47:07,843 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/4150f0b85bebbb7094b7a1d91b0a6f32/.tmp/cf/b85d87029a5646189c7a994b7c99cdba as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/4150f0b85bebbb7094b7a1d91b0a6f32/cf/b85d87029a5646189c7a994b7c99cdba 2025-01-08T10:47:07,848 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/6ff7afa9ac49884297a386018fda40ef/cf/cd90586f90f64d5691980aaf0b4c9136, entries=47, sequenceid=6, filesize=15.3 K 2025-01-08T10:47:07,849 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.HRegion(3040): Finished flush of dataSize ~3.06 KB/3137, heapSize ~6.84 KB/7008, currentSize=0 B/0 for 6ff7afa9ac49884297a386018fda40ef in 60ms, sequenceid=6, compaction requested=false 2025-01-08T10:47:07,849 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportFileSystemStateWithMergeRegion' 2025-01-08T10:47:07,849 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/4150f0b85bebbb7094b7a1d91b0a6f32/cf/b85d87029a5646189c7a994b7c99cdba, entries=3, sequenceid=6, filesize=5.8 K 2025-01-08T10:47:07,850 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.HRegion(2538): Flush status journal for 6ff7afa9ac49884297a386018fda40ef: 2025-01-08T10:47:07,850 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.HRegion(3040): Finished flush of dataSize ~199 B/199, heapSize ~672 B/672, currentSize=0 B/0 for 4150f0b85bebbb7094b7a1d91b0a6f32 in 61ms, sequenceid=6, compaction requested=false 2025-01-08T10:47:07,850 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef. for snaptb0-testExportFileSystemStateWithMergeRegion completed. 2025-01-08T10:47:07,850 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.HRegion(2538): Flush status journal for 4150f0b85bebbb7094b7a1d91b0a6f32: 2025-01-08T10:47:07,850 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32. for snaptb0-testExportFileSystemStateWithMergeRegion completed. 2025-01-08T10:47:07,850 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef.' region-info for snapshot=snaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:07,850 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:47:07,850 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32.' region-info for snapshot=snaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:07,850 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/6ff7afa9ac49884297a386018fda40ef/cf/cd90586f90f64d5691980aaf0b4c9136] hfiles 2025-01-08T10:47:07,850 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/6ff7afa9ac49884297a386018fda40ef/cf/cd90586f90f64d5691980aaf0b4c9136 for snapshot=snaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:07,850 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:47:07,850 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/4150f0b85bebbb7094b7a1d91b0a6f32/cf/b85d87029a5646189c7a994b7c99cdba] hfiles 2025-01-08T10:47:07,850 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/4150f0b85bebbb7094b7a1d91b0a6f32/cf/b85d87029a5646189c7a994b7c99cdba for snapshot=snaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:07,856 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742137_1313 (size=125) 2025-01-08T10:47:07,857 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742137_1313 (size=125) 2025-01-08T10:47:07,857 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742138_1314 (size=125) 2025-01-08T10:47:07,858 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742137_1313 (size=125) 2025-01-08T10:47:07,858 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742138_1314 (size=125) 2025-01-08T10:47:07,858 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef. 2025-01-08T10:47:07,858 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742138_1314 (size=125) 2025-01-08T10:47:07,858 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=109 2025-01-08T10:47:07,858 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32. 2025-01-08T10:47:07,859 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=108 2025-01-08T10:47:07,859 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=109 2025-01-08T10:47:07,859 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportFileSystemStateWithMergeRegion on region 6ff7afa9ac49884297a386018fda40ef 2025-01-08T10:47:07,859 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=109, ppid=107, state=RUNNABLE; SnapshotRegionProcedure 6ff7afa9ac49884297a386018fda40ef 2025-01-08T10:47:07,859 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=108 2025-01-08T10:47:07,859 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportFileSystemStateWithMergeRegion on region 4150f0b85bebbb7094b7a1d91b0a6f32 2025-01-08T10:47:07,859 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=108, ppid=107, state=RUNNABLE; SnapshotRegionProcedure 4150f0b85bebbb7094b7a1d91b0a6f32 2025-01-08T10:47:07,861 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=108, ppid=107, state=SUCCESS; SnapshotRegionProcedure 4150f0b85bebbb7094b7a1d91b0a6f32 in 224 msec 2025-01-08T10:47:07,861 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=109, resume processing ppid=107 2025-01-08T10:47:07,861 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=109, ppid=107, state=SUCCESS; SnapshotRegionProcedure 6ff7afa9ac49884297a386018fda40ef in 224 msec 2025-01-08T10:47:07,861 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T10:47:07,862 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T10:47:07,863 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T10:47:07,863 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T10:47:07,863 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:07,864 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/c4ca4238a0b923820dcc509a6f75849b2025010810301e755fe74f01a4220a74bb7a8296_6ff7afa9ac49884297a386018fda40ef, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/d41d8cd98f00b204e9800998ecf8427e20250108df8d023a67d946d4a1716e125356d5e2_4150f0b85bebbb7094b7a1d91b0a6f32] hfiles 2025-01-08T10:47:07,864 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/c4ca4238a0b923820dcc509a6f75849b2025010810301e755fe74f01a4220a74bb7a8296_6ff7afa9ac49884297a386018fda40ef 2025-01-08T10:47:07,864 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/d41d8cd98f00b204e9800998ecf8427e20250108df8d023a67d946d4a1716e125356d5e2_4150f0b85bebbb7094b7a1d91b0a6f32 2025-01-08T10:47:07,870 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742139_1315 (size=309) 2025-01-08T10:47:07,870 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742139_1315 (size=309) 2025-01-08T10:47:07,871 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742139_1315 (size=309) 2025-01-08T10:47:07,872 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T10:47:07,872 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:07,872 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:07,883 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742140_1316 (size=1023) 2025-01-08T10:47:07,883 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742140_1316 (size=1023) 2025-01-08T10:47:07,884 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742140_1316 (size=1023) 2025-01-08T10:47:07,886 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T10:47:07,891 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T10:47:07,892 DEBUG [PEWorker-3 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemStateWithMergeRegion to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:07,893 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T10:47:07,893 DEBUG [PEWorker-3 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 }, snapshot procedure id = 107 2025-01-08T10:47:07,894 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=107, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } in 270 msec 2025-01-08T10:47:07,930 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=107 2025-01-08T10:47:07,931 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion, procId: 107 completed 2025-01-08T10:47:07,952 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2025-01-08T10:47:07,955 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:46722, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2025-01-08T10:47:07,956 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36283 {}] regionserver.CompactSplit(323): Interrupting running compactions because user switched off compactions 2025-01-08T10:47:07,956 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2025-01-08T10:47:07,958 INFO [RS-EventLoopGroup-5-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:52496, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2025-01-08T10:47:07,958 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38323 {}] regionserver.CompactSplit(323): Interrupting running compactions because user switched off compactions 2025-01-08T10:47:07,958 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2025-01-08T10:47:07,960 INFO [RS-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:58662, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2025-01-08T10:47:07,960 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=45003 {}] regionserver.CompactSplit(323): Interrupting running compactions because user switched off compactions 2025-01-08T10:47:07,962 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.2 create 'testtb-testExportFileSystemStateWithMergeRegion-1', {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T10:47:07,962 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=110, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:07,963 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=110, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T10:47:07,963 DEBUG [PEWorker-2 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:07,963 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "testtb-testExportFileSystemStateWithMergeRegion-1" procId is: 110 2025-01-08T10:47:07,964 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=110, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T10:47:07,964 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=110 2025-01-08T10:47:07,971 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742141_1317 (size=399) 2025-01-08T10:47:07,971 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742141_1317 (size=399) 2025-01-08T10:47:07,971 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742141_1317 (size=399) 2025-01-08T10:47:07,973 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => d6f6f260571a44cef14a139fbdf37112, NAME => 'testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227961.d6f6f260571a44cef14a139fbdf37112.', STARTKEY => '', ENDKEY => '2'}, tableDescriptor='testtb-testExportFileSystemStateWithMergeRegion-1', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:47:07,975 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => 9eb0e33eafa3af0c0eeaba516cde4e33, NAME => 'testtb-testExportFileSystemStateWithMergeRegion-1,2,1736333227961.9eb0e33eafa3af0c0eeaba516cde4e33.', STARTKEY => '2', ENDKEY => ''}, tableDescriptor='testtb-testExportFileSystemStateWithMergeRegion-1', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:47:07,984 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742142_1318 (size=85) 2025-01-08T10:47:07,984 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742142_1318 (size=85) 2025-01-08T10:47:07,984 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742142_1318 (size=85) 2025-01-08T10:47:07,985 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227961.d6f6f260571a44cef14a139fbdf37112.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:47:07,985 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-0 {}] regionserver.HRegion(1681): Closing d6f6f260571a44cef14a139fbdf37112, disabling compactions & flushes 2025-01-08T10:47:07,985 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227961.d6f6f260571a44cef14a139fbdf37112. 2025-01-08T10:47:07,985 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227961.d6f6f260571a44cef14a139fbdf37112. 2025-01-08T10:47:07,985 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227961.d6f6f260571a44cef14a139fbdf37112. after waiting 0 ms 2025-01-08T10:47:07,985 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227961.d6f6f260571a44cef14a139fbdf37112. 2025-01-08T10:47:07,985 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227961.d6f6f260571a44cef14a139fbdf37112. 2025-01-08T10:47:07,985 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-0 {}] regionserver.HRegion(1635): Region close journal for d6f6f260571a44cef14a139fbdf37112: 2025-01-08T10:47:07,986 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742143_1319 (size=85) 2025-01-08T10:47:07,986 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742143_1319 (size=85) 2025-01-08T10:47:07,987 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742143_1319 (size=85) 2025-01-08T10:47:07,987 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithMergeRegion-1,2,1736333227961.9eb0e33eafa3af0c0eeaba516cde4e33.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:47:07,987 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-1 {}] regionserver.HRegion(1681): Closing 9eb0e33eafa3af0c0eeaba516cde4e33, disabling compactions & flushes 2025-01-08T10:47:07,987 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithMergeRegion-1,2,1736333227961.9eb0e33eafa3af0c0eeaba516cde4e33. 2025-01-08T10:47:07,987 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithMergeRegion-1,2,1736333227961.9eb0e33eafa3af0c0eeaba516cde4e33. 2025-01-08T10:47:07,987 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithMergeRegion-1,2,1736333227961.9eb0e33eafa3af0c0eeaba516cde4e33. after waiting 0 ms 2025-01-08T10:47:07,987 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithMergeRegion-1,2,1736333227961.9eb0e33eafa3af0c0eeaba516cde4e33. 2025-01-08T10:47:07,987 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithMergeRegion-1,2,1736333227961.9eb0e33eafa3af0c0eeaba516cde4e33. 2025-01-08T10:47:07,987 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-1 {}] regionserver.HRegion(1635): Region close journal for 9eb0e33eafa3af0c0eeaba516cde4e33: 2025-01-08T10:47:07,988 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=110, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T10:47:07,988 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227961.d6f6f260571a44cef14a139fbdf37112.","families":{"info":[{"qualifier":"regioninfo","vlen":84,"tag":[],"timestamp":"1736333227988"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736333227988"}]},"ts":"1736333227988"} 2025-01-08T10:47:07,989 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportFileSystemStateWithMergeRegion-1,2,1736333227961.9eb0e33eafa3af0c0eeaba516cde4e33.","families":{"info":[{"qualifier":"regioninfo","vlen":84,"tag":[],"timestamp":"1736333227988"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736333227988"}]},"ts":"1736333227988"} 2025-01-08T10:47:07,991 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T10:47:07,992 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=110, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T10:47:07,992 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion-1","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333227992"}]},"ts":"1736333227992"} 2025-01-08T10:47:07,993 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithMergeRegion-1, state=ENABLING in hbase:meta 2025-01-08T10:47:07,997 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(202): Hosts are {67f42ee69b4e=0} racks are {/default-rack=0} 2025-01-08T10:47:07,998 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T10:47:07,998 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T10:47:07,998 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T10:47:07,998 INFO [PEWorker-2 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T10:47:07,998 INFO [PEWorker-2 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T10:47:07,998 INFO [PEWorker-2 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T10:47:07,998 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T10:47:07,998 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=111, ppid=110, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=d6f6f260571a44cef14a139fbdf37112, ASSIGN}, {pid=112, ppid=110, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=9eb0e33eafa3af0c0eeaba516cde4e33, ASSIGN}] 2025-01-08T10:47:07,999 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=112, ppid=110, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=9eb0e33eafa3af0c0eeaba516cde4e33, ASSIGN 2025-01-08T10:47:07,999 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=111, ppid=110, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=d6f6f260571a44cef14a139fbdf37112, ASSIGN 2025-01-08T10:47:08,000 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=112, ppid=110, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=9eb0e33eafa3af0c0eeaba516cde4e33, ASSIGN; state=OFFLINE, location=67f42ee69b4e,36283,1736333108166; forceNewPlan=false, retain=false 2025-01-08T10:47:08,000 INFO [PEWorker-5 {}] assignment.TransitRegionStateProcedure(264): Starting pid=111, ppid=110, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=d6f6f260571a44cef14a139fbdf37112, ASSIGN; state=OFFLINE, location=67f42ee69b4e,38323,1736333108259; forceNewPlan=false, retain=false 2025-01-08T10:47:08,065 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=110 2025-01-08T10:47:08,150 INFO [67f42ee69b4e:34551 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T10:47:08,151 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=112 updating hbase:meta row=9eb0e33eafa3af0c0eeaba516cde4e33, regionState=OPENING, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:08,151 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=111 updating hbase:meta row=d6f6f260571a44cef14a139fbdf37112, regionState=OPENING, regionLocation=67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:08,152 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=113, ppid=112, state=RUNNABLE; OpenRegionProcedure 9eb0e33eafa3af0c0eeaba516cde4e33, server=67f42ee69b4e,36283,1736333108166}] 2025-01-08T10:47:08,153 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=114, ppid=111, state=RUNNABLE; OpenRegionProcedure d6f6f260571a44cef14a139fbdf37112, server=67f42ee69b4e,38323,1736333108259}] 2025-01-08T10:47:08,266 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=110 2025-01-08T10:47:08,304 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:08,304 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:08,307 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] handler.AssignRegionHandler(135): Open testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227961.d6f6f260571a44cef14a139fbdf37112. 2025-01-08T10:47:08,307 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] handler.AssignRegionHandler(135): Open testtb-testExportFileSystemStateWithMergeRegion-1,2,1736333227961.9eb0e33eafa3af0c0eeaba516cde4e33. 2025-01-08T10:47:08,308 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(7285): Opening region: {ENCODED => 9eb0e33eafa3af0c0eeaba516cde4e33, NAME => 'testtb-testExportFileSystemStateWithMergeRegion-1,2,1736333227961.9eb0e33eafa3af0c0eeaba516cde4e33.', STARTKEY => '2', ENDKEY => ''} 2025-01-08T10:47:08,308 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(7285): Opening region: {ENCODED => d6f6f260571a44cef14a139fbdf37112, NAME => 'testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227961.d6f6f260571a44cef14a139fbdf37112.', STARTKEY => '', ENDKEY => '2'} 2025-01-08T10:47:08,308 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227961.d6f6f260571a44cef14a139fbdf37112. service=AccessControlService 2025-01-08T10:47:08,308 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportFileSystemStateWithMergeRegion-1,2,1736333227961.9eb0e33eafa3af0c0eeaba516cde4e33. service=AccessControlService 2025-01-08T10:47:08,308 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:47:08,308 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:47:08,308 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportFileSystemStateWithMergeRegion-1 d6f6f260571a44cef14a139fbdf37112 2025-01-08T10:47:08,308 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportFileSystemStateWithMergeRegion-1 9eb0e33eafa3af0c0eeaba516cde4e33 2025-01-08T10:47:08,308 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithMergeRegion-1,2,1736333227961.9eb0e33eafa3af0c0eeaba516cde4e33.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:47:08,308 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227961.d6f6f260571a44cef14a139fbdf37112.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:47:08,308 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(7327): checking encryption for 9eb0e33eafa3af0c0eeaba516cde4e33 2025-01-08T10:47:08,308 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(7327): checking encryption for d6f6f260571a44cef14a139fbdf37112 2025-01-08T10:47:08,308 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(7330): checking classloading for 9eb0e33eafa3af0c0eeaba516cde4e33 2025-01-08T10:47:08,309 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(7330): checking classloading for d6f6f260571a44cef14a139fbdf37112 2025-01-08T10:47:08,310 INFO [StoreOpener-9eb0e33eafa3af0c0eeaba516cde4e33-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 9eb0e33eafa3af0c0eeaba516cde4e33 2025-01-08T10:47:08,310 INFO [StoreOpener-d6f6f260571a44cef14a139fbdf37112-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region d6f6f260571a44cef14a139fbdf37112 2025-01-08T10:47:08,311 INFO [StoreOpener-9eb0e33eafa3af0c0eeaba516cde4e33-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 9eb0e33eafa3af0c0eeaba516cde4e33 columnFamilyName cf 2025-01-08T10:47:08,311 INFO [StoreOpener-d6f6f260571a44cef14a139fbdf37112-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region d6f6f260571a44cef14a139fbdf37112 columnFamilyName cf 2025-01-08T10:47:08,311 DEBUG [StoreOpener-9eb0e33eafa3af0c0eeaba516cde4e33-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:08,311 DEBUG [StoreOpener-d6f6f260571a44cef14a139fbdf37112-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:08,312 INFO [StoreOpener-9eb0e33eafa3af0c0eeaba516cde4e33-1 {}] regionserver.HStore(327): Store=9eb0e33eafa3af0c0eeaba516cde4e33/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:47:08,312 INFO [StoreOpener-d6f6f260571a44cef14a139fbdf37112-1 {}] regionserver.HStore(327): Store=d6f6f260571a44cef14a139fbdf37112/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:47:08,312 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/d6f6f260571a44cef14a139fbdf37112 2025-01-08T10:47:08,312 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/9eb0e33eafa3af0c0eeaba516cde4e33 2025-01-08T10:47:08,313 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/d6f6f260571a44cef14a139fbdf37112 2025-01-08T10:47:08,313 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/9eb0e33eafa3af0c0eeaba516cde4e33 2025-01-08T10:47:08,315 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(1085): writing seq id for 9eb0e33eafa3af0c0eeaba516cde4e33 2025-01-08T10:47:08,315 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(1085): writing seq id for d6f6f260571a44cef14a139fbdf37112 2025-01-08T10:47:08,316 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/9eb0e33eafa3af0c0eeaba516cde4e33/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:47:08,317 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/d6f6f260571a44cef14a139fbdf37112/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:47:08,317 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(1102): Opened d6f6f260571a44cef14a139fbdf37112; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=70972097, jitterRate=0.05756665766239166}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T10:47:08,317 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(1102): Opened 9eb0e33eafa3af0c0eeaba516cde4e33; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=60065729, jitterRate=-0.10495088994503021}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T10:47:08,318 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(1001): Region open journal for d6f6f260571a44cef14a139fbdf37112: 2025-01-08T10:47:08,318 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(1001): Region open journal for 9eb0e33eafa3af0c0eeaba516cde4e33: 2025-01-08T10:47:08,318 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227961.d6f6f260571a44cef14a139fbdf37112., pid=114, masterSystemTime=1736333228304 2025-01-08T10:47:08,319 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportFileSystemStateWithMergeRegion-1,2,1736333227961.9eb0e33eafa3af0c0eeaba516cde4e33., pid=113, masterSystemTime=1736333228304 2025-01-08T10:47:08,320 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227961.d6f6f260571a44cef14a139fbdf37112. 2025-01-08T10:47:08,320 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] handler.AssignRegionHandler(164): Opened testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227961.d6f6f260571a44cef14a139fbdf37112. 2025-01-08T10:47:08,320 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=111 updating hbase:meta row=d6f6f260571a44cef14a139fbdf37112, regionState=OPEN, openSeqNum=2, regionLocation=67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:08,320 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportFileSystemStateWithMergeRegion-1,2,1736333227961.9eb0e33eafa3af0c0eeaba516cde4e33. 2025-01-08T10:47:08,320 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] handler.AssignRegionHandler(164): Opened testtb-testExportFileSystemStateWithMergeRegion-1,2,1736333227961.9eb0e33eafa3af0c0eeaba516cde4e33. 2025-01-08T10:47:08,322 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=112 updating hbase:meta row=9eb0e33eafa3af0c0eeaba516cde4e33, regionState=OPEN, openSeqNum=2, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:08,324 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=114, resume processing ppid=111 2025-01-08T10:47:08,324 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=114, ppid=111, state=SUCCESS; OpenRegionProcedure d6f6f260571a44cef14a139fbdf37112, server=67f42ee69b4e,38323,1736333108259 in 170 msec 2025-01-08T10:47:08,325 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=113, resume processing ppid=112 2025-01-08T10:47:08,325 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=113, ppid=112, state=SUCCESS; OpenRegionProcedure 9eb0e33eafa3af0c0eeaba516cde4e33, server=67f42ee69b4e,36283,1736333108166 in 172 msec 2025-01-08T10:47:08,326 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=111, ppid=110, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=d6f6f260571a44cef14a139fbdf37112, ASSIGN in 326 msec 2025-01-08T10:47:08,326 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=112, resume processing ppid=110 2025-01-08T10:47:08,327 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=112, ppid=110, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=9eb0e33eafa3af0c0eeaba516cde4e33, ASSIGN in 327 msec 2025-01-08T10:47:08,327 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=110, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T10:47:08,327 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion-1","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333228327"}]},"ts":"1736333228327"} 2025-01-08T10:47:08,328 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithMergeRegion-1, state=ENABLED in hbase:meta 2025-01-08T10:47:08,330 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=110, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T10:47:08,331 DEBUG [PEWorker-4 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testExportFileSystemStateWithMergeRegion-1 jenkins: RWXCA 2025-01-08T10:47:08,332 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36283 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemStateWithMergeRegion-1], kv [jenkins: RWXCA] 2025-01-08T10:47:08,334 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:08,334 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:08,334 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:08,334 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:08,336 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:08,336 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:08,336 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:08,336 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion-1 with data PBUF\x0AW\x0A\x07jenkins\x12L\x08\x03"H\x0A<\x0A\x07default\x121testtb-testExportFileSystemStateWithMergeRegion-1 \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:08,336 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion-1 with data PBUF\x0AW\x0A\x07jenkins\x12L\x08\x03"H\x0A<\x0A\x07default\x121testtb-testExportFileSystemStateWithMergeRegion-1 \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:08,336 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion-1 with data PBUF\x0AW\x0A\x07jenkins\x12L\x08\x03"H\x0A<\x0A\x07default\x121testtb-testExportFileSystemStateWithMergeRegion-1 \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:08,337 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:08,337 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion-1 with data PBUF\x0AW\x0A\x07jenkins\x12L\x08\x03"H\x0A<\x0A\x07default\x121testtb-testExportFileSystemStateWithMergeRegion-1 \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:08,338 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=110, state=SUCCESS; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 in 374 msec 2025-01-08T10:47:08,567 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=110 2025-01-08T10:47:08,568 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion-1, procId: 110 completed 2025-01-08T10:47:08,588 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster$2(2234): Client=jenkins//172.17.0.2 merge regions [d6f6f260571a44cef14a139fbdf37112, 9eb0e33eafa3af0c0eeaba516cde4e33] 2025-01-08T10:47:08,593 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=115, state=RUNNABLE:MERGE_TABLE_REGIONS_PREPARE; MergeTableRegionsProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, regions=[d6f6f260571a44cef14a139fbdf37112, 9eb0e33eafa3af0c0eeaba516cde4e33], force=true 2025-01-08T10:47:08,594 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=115, state=RUNNABLE:MERGE_TABLE_REGIONS_PREPARE; MergeTableRegionsProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, regions=[d6f6f260571a44cef14a139fbdf37112, 9eb0e33eafa3af0c0eeaba516cde4e33], force=true 2025-01-08T10:47:08,594 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=115, state=RUNNABLE:MERGE_TABLE_REGIONS_PREPARE; MergeTableRegionsProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, regions=[d6f6f260571a44cef14a139fbdf37112, 9eb0e33eafa3af0c0eeaba516cde4e33], force=true 2025-01-08T10:47:08,594 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=115, state=RUNNABLE:MERGE_TABLE_REGIONS_PREPARE; MergeTableRegionsProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, regions=[d6f6f260571a44cef14a139fbdf37112, 9eb0e33eafa3af0c0eeaba516cde4e33], force=true 2025-01-08T10:47:08,596 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=115 2025-01-08T10:47:08,604 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=116, ppid=115, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=d6f6f260571a44cef14a139fbdf37112, UNASSIGN}, {pid=117, ppid=115, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=9eb0e33eafa3af0c0eeaba516cde4e33, UNASSIGN}] 2025-01-08T10:47:08,605 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=117, ppid=115, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=9eb0e33eafa3af0c0eeaba516cde4e33, UNASSIGN 2025-01-08T10:47:08,605 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=116, ppid=115, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=d6f6f260571a44cef14a139fbdf37112, UNASSIGN 2025-01-08T10:47:08,606 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=117 updating hbase:meta row=9eb0e33eafa3af0c0eeaba516cde4e33, regionState=CLOSING, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:08,606 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=116 updating hbase:meta row=d6f6f260571a44cef14a139fbdf37112, regionState=CLOSING, regionLocation=67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:08,607 DEBUG [PEWorker-3 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: true: evictOnSplit: true: evictOnClose: false 2025-01-08T10:47:08,607 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=118, ppid=117, state=RUNNABLE; CloseRegionProcedure 9eb0e33eafa3af0c0eeaba516cde4e33, server=67f42ee69b4e,36283,1736333108166}] 2025-01-08T10:47:08,608 DEBUG [PEWorker-5 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: true: evictOnSplit: true: evictOnClose: false 2025-01-08T10:47:08,608 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=119, ppid=116, state=RUNNABLE; CloseRegionProcedure d6f6f260571a44cef14a139fbdf37112, server=67f42ee69b4e,38323,1736333108259}] 2025-01-08T10:47:08,697 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=115 2025-01-08T10:47:08,759 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:08,759 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:08,760 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] handler.UnassignRegionHandler(124): Close 9eb0e33eafa3af0c0eeaba516cde4e33 2025-01-08T10:47:08,760 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] handler.UnassignRegionHandler(124): Close d6f6f260571a44cef14a139fbdf37112 2025-01-08T10:47:08,760 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] handler.UnassignRegionHandler(138): Unassign region: split region: true: evictCache: true 2025-01-08T10:47:08,760 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] handler.UnassignRegionHandler(138): Unassign region: split region: true: evictCache: true 2025-01-08T10:47:08,760 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegion(1681): Closing d6f6f260571a44cef14a139fbdf37112, disabling compactions & flushes 2025-01-08T10:47:08,760 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegion(1681): Closing 9eb0e33eafa3af0c0eeaba516cde4e33, disabling compactions & flushes 2025-01-08T10:47:08,760 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227961.d6f6f260571a44cef14a139fbdf37112. 2025-01-08T10:47:08,760 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227961.d6f6f260571a44cef14a139fbdf37112. 2025-01-08T10:47:08,760 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithMergeRegion-1,2,1736333227961.9eb0e33eafa3af0c0eeaba516cde4e33. 2025-01-08T10:47:08,760 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227961.d6f6f260571a44cef14a139fbdf37112. after waiting 0 ms 2025-01-08T10:47:08,760 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithMergeRegion-1,2,1736333227961.9eb0e33eafa3af0c0eeaba516cde4e33. 2025-01-08T10:47:08,760 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227961.d6f6f260571a44cef14a139fbdf37112. 2025-01-08T10:47:08,760 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithMergeRegion-1,2,1736333227961.9eb0e33eafa3af0c0eeaba516cde4e33. after waiting 0 ms 2025-01-08T10:47:08,760 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegion(2837): Flushing d6f6f260571a44cef14a139fbdf37112 1/1 column families, dataSize=24 B heapSize=352 B 2025-01-08T10:47:08,760 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithMergeRegion-1,2,1736333227961.9eb0e33eafa3af0c0eeaba516cde4e33. 2025-01-08T10:47:08,761 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegion(2837): Flushing 9eb0e33eafa3af0c0eeaba516cde4e33 1/1 column families, dataSize=24 B heapSize=352 B 2025-01-08T10:47:08,778 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/9eb0e33eafa3af0c0eeaba516cde4e33/.tmp/cf/fe11a77ddadb4b4898e3d882f5f4d1bd is 28, key is 2/cf:/1736333228575/Put/seqid=0 2025-01-08T10:47:08,783 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/d6f6f260571a44cef14a139fbdf37112/.tmp/cf/a5e18dbe9aa04fce8b1fb88298e33f9a is 28, key is 1/cf:/1736333228571/Put/seqid=0 2025-01-08T10:47:08,785 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742144_1320 (size=4945) 2025-01-08T10:47:08,785 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742144_1320 (size=4945) 2025-01-08T10:47:08,786 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742144_1320 (size=4945) 2025-01-08T10:47:08,786 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=24 B at sequenceid=5 (bloomFilter=true), to=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/9eb0e33eafa3af0c0eeaba516cde4e33/.tmp/cf/fe11a77ddadb4b4898e3d882f5f4d1bd 2025-01-08T10:47:08,792 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/9eb0e33eafa3af0c0eeaba516cde4e33/.tmp/cf/fe11a77ddadb4b4898e3d882f5f4d1bd as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/9eb0e33eafa3af0c0eeaba516cde4e33/cf/fe11a77ddadb4b4898e3d882f5f4d1bd 2025-01-08T10:47:08,796 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/9eb0e33eafa3af0c0eeaba516cde4e33/cf/fe11a77ddadb4b4898e3d882f5f4d1bd, entries=1, sequenceid=5, filesize=4.8 K 2025-01-08T10:47:08,797 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegion(3040): Finished flush of dataSize ~24 B/24, heapSize ~336 B/336, currentSize=0 B/0 for 9eb0e33eafa3af0c0eeaba516cde4e33 in 37ms, sequenceid=5, compaction requested=false 2025-01-08T10:47:08,797 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportFileSystemStateWithMergeRegion-1' 2025-01-08T10:47:08,806 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742145_1321 (size=4945) 2025-01-08T10:47:08,806 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742145_1321 (size=4945) 2025-01-08T10:47:08,807 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742145_1321 (size=4945) 2025-01-08T10:47:08,807 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=24 B at sequenceid=5 (bloomFilter=true), to=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/d6f6f260571a44cef14a139fbdf37112/.tmp/cf/a5e18dbe9aa04fce8b1fb88298e33f9a 2025-01-08T10:47:08,811 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/9eb0e33eafa3af0c0eeaba516cde4e33/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2025-01-08T10:47:08,812 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:47:08,812 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/d6f6f260571a44cef14a139fbdf37112/.tmp/cf/a5e18dbe9aa04fce8b1fb88298e33f9a as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/d6f6f260571a44cef14a139fbdf37112/cf/a5e18dbe9aa04fce8b1fb88298e33f9a 2025-01-08T10:47:08,812 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithMergeRegion-1,2,1736333227961.9eb0e33eafa3af0c0eeaba516cde4e33. 2025-01-08T10:47:08,812 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegion(1635): Region close journal for 9eb0e33eafa3af0c0eeaba516cde4e33: 2025-01-08T10:47:08,814 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] handler.UnassignRegionHandler(170): Closed 9eb0e33eafa3af0c0eeaba516cde4e33 2025-01-08T10:47:08,814 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=117 updating hbase:meta row=9eb0e33eafa3af0c0eeaba516cde4e33, regionState=CLOSED 2025-01-08T10:47:08,818 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=118, resume processing ppid=117 2025-01-08T10:47:08,818 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=118, ppid=117, state=SUCCESS; CloseRegionProcedure 9eb0e33eafa3af0c0eeaba516cde4e33, server=67f42ee69b4e,36283,1736333108166 in 209 msec 2025-01-08T10:47:08,819 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=117, ppid=115, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=9eb0e33eafa3af0c0eeaba516cde4e33, UNASSIGN in 214 msec 2025-01-08T10:47:08,824 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/d6f6f260571a44cef14a139fbdf37112/cf/a5e18dbe9aa04fce8b1fb88298e33f9a, entries=1, sequenceid=5, filesize=4.8 K 2025-01-08T10:47:08,825 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegion(3040): Finished flush of dataSize ~24 B/24, heapSize ~336 B/336, currentSize=0 B/0 for d6f6f260571a44cef14a139fbdf37112 in 65ms, sequenceid=5, compaction requested=false 2025-01-08T10:47:08,829 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/d6f6f260571a44cef14a139fbdf37112/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2025-01-08T10:47:08,829 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:47:08,829 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227961.d6f6f260571a44cef14a139fbdf37112. 2025-01-08T10:47:08,829 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegion(1635): Region close journal for d6f6f260571a44cef14a139fbdf37112: 2025-01-08T10:47:08,831 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] handler.UnassignRegionHandler(170): Closed d6f6f260571a44cef14a139fbdf37112 2025-01-08T10:47:08,831 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=116 updating hbase:meta row=d6f6f260571a44cef14a139fbdf37112, regionState=CLOSED 2025-01-08T10:47:08,834 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=119, resume processing ppid=116 2025-01-08T10:47:08,834 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=119, ppid=116, state=SUCCESS; CloseRegionProcedure d6f6f260571a44cef14a139fbdf37112, server=67f42ee69b4e,38323,1736333108259 in 224 msec 2025-01-08T10:47:08,835 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=116, resume processing ppid=115 2025-01-08T10:47:08,836 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=116, ppid=115, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=d6f6f260571a44cef14a139fbdf37112, UNASSIGN in 230 msec 2025-01-08T10:47:08,849 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742146_1322 (size=84) 2025-01-08T10:47:08,849 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742146_1322 (size=84) 2025-01-08T10:47:08,849 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742146_1322 (size=84) 2025-01-08T10:47:08,851 DEBUG [PEWorker-4 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:08,865 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742147_1323 (size=20) 2025-01-08T10:47:08,865 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742147_1323 (size=20) 2025-01-08T10:47:08,866 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742147_1323 (size=20) 2025-01-08T10:47:08,868 DEBUG [PEWorker-4 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:08,874 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742148_1324 (size=21) 2025-01-08T10:47:08,874 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742148_1324 (size=21) 2025-01-08T10:47:08,874 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742148_1324 (size=21) 2025-01-08T10:47:08,880 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742149_1325 (size=84) 2025-01-08T10:47:08,881 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742149_1325 (size=84) 2025-01-08T10:47:08,881 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742149_1325 (size=84) 2025-01-08T10:47:08,882 DEBUG [PEWorker-4 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:08,896 DEBUG [PEWorker-4 {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/96d4bb39d5de2c0741f205ac02281c75/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=-1 2025-01-08T10:47:08,898 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227961.d6f6f260571a44cef14a139fbdf37112.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"9223372036854775807"}]},"ts":"9223372036854775807"} 2025-01-08T10:47:08,898 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion-1,2,1736333227961.9eb0e33eafa3af0c0eeaba516cde4e33.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"9223372036854775807"}]},"ts":"9223372036854775807"} 2025-01-08T10:47:08,898 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":7,"row":"testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227962.96d4bb39d5de2c0741f205ac02281c75.","families":{"info":[{"qualifier":"regioninfo","vlen":83,"tag":[],"timestamp":"9223372036854775807"},{"qualifier":"merge0000","vlen":84,"tag":[],"timestamp":"9223372036854775807"},{"qualifier":"merge0001","vlen":84,"tag":[],"timestamp":"9223372036854775807"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"9223372036854775807"}]},"ts":"9223372036854775807"} 2025-01-08T10:47:08,898 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=115 2025-01-08T10:47:08,927 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=120, ppid=115, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=96d4bb39d5de2c0741f205ac02281c75, ASSIGN}] 2025-01-08T10:47:08,928 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=120, ppid=115, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=96d4bb39d5de2c0741f205ac02281c75, ASSIGN 2025-01-08T10:47:08,929 INFO [PEWorker-1 {}] assignment.TransitRegionStateProcedure(264): Starting pid=120, ppid=115, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=96d4bb39d5de2c0741f205ac02281c75, ASSIGN; state=MERGED, location=67f42ee69b4e,38323,1736333108259; forceNewPlan=false, retain=false 2025-01-08T10:47:09,079 INFO [67f42ee69b4e:34551 {}] balancer.BaseLoadBalancer(546): Reassigned 1 regions. 1 retained the pre-restart assignment. 2025-01-08T10:47:09,079 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=120 updating hbase:meta row=96d4bb39d5de2c0741f205ac02281c75, regionState=OPENING, regionLocation=67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:09,082 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=121, ppid=120, state=RUNNABLE; OpenRegionProcedure 96d4bb39d5de2c0741f205ac02281c75, server=67f42ee69b4e,38323,1736333108259}] 2025-01-08T10:47:09,128 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0005_000001 (auth:SIMPLE) from 127.0.0.1:39798 2025-01-08T10:47:09,140 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_2/usercache/jenkins/appcache/application_1736333116683_0005/container_1736333116683_0005_01_000001/launch_container.sh] 2025-01-08T10:47:09,140 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_2/usercache/jenkins/appcache/application_1736333116683_0005/container_1736333116683_0005_01_000001/container_tokens] 2025-01-08T10:47:09,140 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_2/usercache/jenkins/appcache/application_1736333116683_0005/container_1736333116683_0005_01_000001/sysfs] 2025-01-08T10:47:09,199 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=115 2025-01-08T10:47:09,233 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:09,236 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] handler.AssignRegionHandler(135): Open testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227962.96d4bb39d5de2c0741f205ac02281c75. 2025-01-08T10:47:09,236 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(7285): Opening region: {ENCODED => 96d4bb39d5de2c0741f205ac02281c75, NAME => 'testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227962.96d4bb39d5de2c0741f205ac02281c75.', STARTKEY => '', ENDKEY => ''} 2025-01-08T10:47:09,237 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227962.96d4bb39d5de2c0741f205ac02281c75. service=AccessControlService 2025-01-08T10:47:09,237 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:47:09,237 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportFileSystemStateWithMergeRegion-1 96d4bb39d5de2c0741f205ac02281c75 2025-01-08T10:47:09,237 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227962.96d4bb39d5de2c0741f205ac02281c75.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:47:09,237 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(7327): checking encryption for 96d4bb39d5de2c0741f205ac02281c75 2025-01-08T10:47:09,237 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(7330): checking classloading for 96d4bb39d5de2c0741f205ac02281c75 2025-01-08T10:47:09,239 INFO [StoreOpener-96d4bb39d5de2c0741f205ac02281c75-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 96d4bb39d5de2c0741f205ac02281c75 2025-01-08T10:47:09,239 INFO [StoreOpener-96d4bb39d5de2c0741f205ac02281c75-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 96d4bb39d5de2c0741f205ac02281c75 columnFamilyName cf 2025-01-08T10:47:09,239 DEBUG [StoreOpener-96d4bb39d5de2c0741f205ac02281c75-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:09,259 DEBUG [StoreOpener-96d4bb39d5de2c0741f205ac02281c75-1 {}] regionserver.StoreEngine(277): loaded hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/96d4bb39d5de2c0741f205ac02281c75/cf/a5e18dbe9aa04fce8b1fb88298e33f9a.d6f6f260571a44cef14a139fbdf37112->hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/d6f6f260571a44cef14a139fbdf37112/cf/a5e18dbe9aa04fce8b1fb88298e33f9a-top 2025-01-08T10:47:09,263 DEBUG [StoreOpener-96d4bb39d5de2c0741f205ac02281c75-1 {}] regionserver.StoreEngine(277): loaded hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/96d4bb39d5de2c0741f205ac02281c75/cf/fe11a77ddadb4b4898e3d882f5f4d1bd.9eb0e33eafa3af0c0eeaba516cde4e33->hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/9eb0e33eafa3af0c0eeaba516cde4e33/cf/fe11a77ddadb4b4898e3d882f5f4d1bd-top 2025-01-08T10:47:09,264 INFO [StoreOpener-96d4bb39d5de2c0741f205ac02281c75-1 {}] regionserver.HStore(327): Store=96d4bb39d5de2c0741f205ac02281c75/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:47:09,265 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/96d4bb39d5de2c0741f205ac02281c75 2025-01-08T10:47:09,266 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/96d4bb39d5de2c0741f205ac02281c75 2025-01-08T10:47:09,268 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(1085): writing seq id for 96d4bb39d5de2c0741f205ac02281c75 2025-01-08T10:47:09,269 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(1102): Opened 96d4bb39d5de2c0741f205ac02281c75; next sequenceid=9; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=67556667, jitterRate=0.006672784686088562}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T10:47:09,269 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(1001): Region open journal for 96d4bb39d5de2c0741f205ac02281c75: 2025-01-08T10:47:09,270 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227962.96d4bb39d5de2c0741f205ac02281c75., pid=121, masterSystemTime=1736333229233 2025-01-08T10:47:09,271 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.CompactSplit(342): Ignoring compaction request for testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227962.96d4bb39d5de2c0741f205ac02281c75.,because compaction is disabled. 2025-01-08T10:47:09,272 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227962.96d4bb39d5de2c0741f205ac02281c75. 2025-01-08T10:47:09,272 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] handler.AssignRegionHandler(164): Opened testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227962.96d4bb39d5de2c0741f205ac02281c75. 2025-01-08T10:47:09,273 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=120 updating hbase:meta row=96d4bb39d5de2c0741f205ac02281c75, regionState=OPEN, openSeqNum=9, regionLocation=67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:09,279 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=121, resume processing ppid=120 2025-01-08T10:47:09,279 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=121, ppid=120, state=SUCCESS; OpenRegionProcedure 96d4bb39d5de2c0741f205ac02281c75, server=67f42ee69b4e,38323,1736333108259 in 195 msec 2025-01-08T10:47:09,283 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=120, resume processing ppid=115 2025-01-08T10:47:09,283 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=120, ppid=115, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=96d4bb39d5de2c0741f205ac02281c75, ASSIGN in 352 msec 2025-01-08T10:47:09,284 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=115, state=SUCCESS; MergeTableRegionsProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, regions=[d6f6f260571a44cef14a139fbdf37112, 9eb0e33eafa3af0c0eeaba516cde4e33], force=true in 693 msec 2025-01-08T10:47:09,700 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=115 2025-01-08T10:47:09,700 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: MERGE_REGIONS, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion-1, procId: 115 completed 2025-01-08T10:47:09,701 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.2 snapshot request for:{ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } 2025-01-08T10:47:09,701 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736333229701 (current time:1736333229701). 2025-01-08T10:47:09,701 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T10:47:09,701 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testExportFileSystemStateWithMergeRegion-1 VERSION not specified, setting to 2 2025-01-08T10:47:09,701 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T10:47:09,702 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x50c2e069 to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@467e857f 2025-01-08T10:47:09,705 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2889162c, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:47:09,707 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:47:09,708 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:46728, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:47:09,709 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x50c2e069 to 127.0.0.1:52367 2025-01-08T10:47:09,709 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:47:09,710 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x128cb9ff to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@206b83ac 2025-01-08T10:47:09,714 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4033c9fb, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:47:09,716 DEBUG [hconnection-0x43eead50-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:47:09,717 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:46738, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:47:09,720 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x128cb9ff to 127.0.0.1:52367 2025-01-08T10:47:09,720 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:47:09,720 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemStateWithMergeRegion-1], kv [jenkins: RWXCA] 2025-01-08T10:47:09,721 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T10:47:09,723 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=122, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } 2025-01-08T10:47:09,723 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 }, snapshot procedure id = 122 2025-01-08T10:47:09,724 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=122 2025-01-08T10:47:09,726 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T10:47:09,727 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T10:47:09,729 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T10:47:09,744 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742150_1326 (size=216) 2025-01-08T10:47:09,744 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742150_1326 (size=216) 2025-01-08T10:47:09,745 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742150_1326 (size=216) 2025-01-08T10:47:09,746 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T10:47:09,746 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=123, ppid=122, state=RUNNABLE; SnapshotRegionProcedure 96d4bb39d5de2c0741f205ac02281c75}] 2025-01-08T10:47:09,747 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=123, ppid=122, state=RUNNABLE; SnapshotRegionProcedure 96d4bb39d5de2c0741f205ac02281c75 2025-01-08T10:47:09,827 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=122 2025-01-08T10:47:09,898 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:09,899 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=38323 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=123 2025-01-08T10:47:09,899 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227962.96d4bb39d5de2c0741f205ac02281c75. 2025-01-08T10:47:09,899 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] regionserver.HRegion(2538): Flush status journal for 96d4bb39d5de2c0741f205ac02281c75: 2025-01-08T10:47:09,899 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227962.96d4bb39d5de2c0741f205ac02281c75. for snaptb0-testExportFileSystemStateWithMergeRegion-1 completed. 2025-01-08T10:47:09,899 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227962.96d4bb39d5de2c0741f205ac02281c75.' region-info for snapshot=snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:09,900 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:47:09,900 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/96d4bb39d5de2c0741f205ac02281c75/cf/a5e18dbe9aa04fce8b1fb88298e33f9a.d6f6f260571a44cef14a139fbdf37112->hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/d6f6f260571a44cef14a139fbdf37112/cf/a5e18dbe9aa04fce8b1fb88298e33f9a-top, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/96d4bb39d5de2c0741f205ac02281c75/cf/fe11a77ddadb4b4898e3d882f5f4d1bd.9eb0e33eafa3af0c0eeaba516cde4e33->hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/9eb0e33eafa3af0c0eeaba516cde4e33/cf/fe11a77ddadb4b4898e3d882f5f4d1bd-top] hfiles 2025-01-08T10:47:09,900 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] snapshot.SnapshotManifest(265): Adding reference for file (1/2): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/96d4bb39d5de2c0741f205ac02281c75/cf/a5e18dbe9aa04fce8b1fb88298e33f9a.d6f6f260571a44cef14a139fbdf37112 for snapshot=snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:09,901 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] snapshot.SnapshotManifest(265): Adding reference for file (2/2): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/96d4bb39d5de2c0741f205ac02281c75/cf/fe11a77ddadb4b4898e3d882f5f4d1bd.9eb0e33eafa3af0c0eeaba516cde4e33 for snapshot=snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:09,911 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742151_1327 (size=269) 2025-01-08T10:47:09,911 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742151_1327 (size=269) 2025-01-08T10:47:09,911 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742151_1327 (size=269) 2025-01-08T10:47:09,912 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227962.96d4bb39d5de2c0741f205ac02281c75. 2025-01-08T10:47:09,912 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=123 2025-01-08T10:47:09,912 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=123 2025-01-08T10:47:09,912 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportFileSystemStateWithMergeRegion-1 on region 96d4bb39d5de2c0741f205ac02281c75 2025-01-08T10:47:09,913 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=123, ppid=122, state=RUNNABLE; SnapshotRegionProcedure 96d4bb39d5de2c0741f205ac02281c75 2025-01-08T10:47:09,915 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=123, resume processing ppid=122 2025-01-08T10:47:09,915 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=123, ppid=122, state=SUCCESS; SnapshotRegionProcedure 96d4bb39d5de2c0741f205ac02281c75 in 167 msec 2025-01-08T10:47:09,915 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T10:47:09,916 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T10:47:09,916 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T10:47:09,917 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:09,917 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:09,926 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742152_1328 (size=670) 2025-01-08T10:47:09,927 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742152_1328 (size=670) 2025-01-08T10:47:09,927 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742152_1328 (size=670) 2025-01-08T10:47:09,930 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T10:47:09,936 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T10:47:09,936 DEBUG [PEWorker-1 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemStateWithMergeRegion-1 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:09,938 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T10:47:09,938 DEBUG [PEWorker-1 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 }, snapshot procedure id = 122 2025-01-08T10:47:09,940 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=122, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } in 217 msec 2025-01-08T10:47:10,029 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=122 2025-01-08T10:47:10,029 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion-1, procId: 122 completed 2025-01-08T10:47:10,029 INFO [Time-limited test {}] snapshot.TestExportSnapshot(468): HDFS export destination path: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333230029 2025-01-08T10:47:10,029 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=hdfs://localhost:37495, tgtDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333230029, rawTgtDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333230029, srcFsUri=hdfs://localhost:37495, srcDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:47:10,086 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:37495, inputRoot=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:47:10,086 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-832679797_22, ugi=jenkins (auth:SIMPLE)]], outputRoot=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333230029, skipTmp=false, initialOutputSnapshotDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333230029/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:10,088 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T10:47:10,094 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333230029/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:10,163 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742153_1329 (size=216) 2025-01-08T10:47:10,163 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742153_1329 (size=216) 2025-01-08T10:47:10,163 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742153_1329 (size=216) 2025-01-08T10:47:10,196 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742154_1330 (size=670) 2025-01-08T10:47:10,196 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742154_1330 (size=670) 2025-01-08T10:47:10,196 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742154_1330 (size=670) 2025-01-08T10:47:10,460 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T10:47:10,760 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-13065403786959587877.jar 2025-01-08T10:47:10,760 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:10,761 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:10,761 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:12,147 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-6649946254869675606.jar 2025-01-08T10:47:12,148 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:12,149 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:12,252 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-17010337373556119715.jar 2025-01-08T10:47:12,253 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:12,253 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:12,254 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:12,254 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:12,255 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:12,255 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:12,255 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T10:47:12,256 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T10:47:12,256 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T10:47:12,257 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T10:47:12,257 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T10:47:12,257 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T10:47:12,258 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T10:47:12,258 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T10:47:12,259 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T10:47:12,259 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T10:47:12,259 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T10:47:12,260 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T10:47:12,260 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:47:12,261 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:47:12,261 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:47:12,261 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:47:12,262 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:47:12,262 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:47:12,263 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:47:12,425 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742155_1331 (size=127628) 2025-01-08T10:47:12,426 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742155_1331 (size=127628) 2025-01-08T10:47:12,428 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742155_1331 (size=127628) 2025-01-08T10:47:12,471 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742156_1332 (size=2172137) 2025-01-08T10:47:12,472 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742156_1332 (size=2172137) 2025-01-08T10:47:12,475 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742156_1332 (size=2172137) 2025-01-08T10:47:12,541 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742157_1333 (size=213228) 2025-01-08T10:47:12,541 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742157_1333 (size=213228) 2025-01-08T10:47:12,542 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742157_1333 (size=213228) 2025-01-08T10:47:12,576 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742158_1334 (size=1877034) 2025-01-08T10:47:12,585 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742158_1334 (size=1877034) 2025-01-08T10:47:12,586 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742158_1334 (size=1877034) 2025-01-08T10:47:12,631 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742159_1335 (size=533455) 2025-01-08T10:47:12,632 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742159_1335 (size=533455) 2025-01-08T10:47:12,633 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742159_1335 (size=533455) 2025-01-08T10:47:12,709 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742160_1336 (size=7280644) 2025-01-08T10:47:12,709 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742160_1336 (size=7280644) 2025-01-08T10:47:12,709 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742160_1336 (size=7280644) 2025-01-08T10:47:12,747 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742161_1337 (size=4188619) 2025-01-08T10:47:12,748 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742161_1337 (size=4188619) 2025-01-08T10:47:12,749 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742161_1337 (size=4188619) 2025-01-08T10:47:12,772 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742162_1338 (size=20406) 2025-01-08T10:47:12,772 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742162_1338 (size=20406) 2025-01-08T10:47:12,773 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742162_1338 (size=20406) 2025-01-08T10:47:12,788 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742163_1339 (size=75495) 2025-01-08T10:47:12,788 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742163_1339 (size=75495) 2025-01-08T10:47:12,789 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742163_1339 (size=75495) 2025-01-08T10:47:12,806 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742164_1340 (size=45609) 2025-01-08T10:47:12,806 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742164_1340 (size=45609) 2025-01-08T10:47:12,806 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742164_1340 (size=45609) 2025-01-08T10:47:12,836 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742165_1341 (size=110084) 2025-01-08T10:47:12,837 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742165_1341 (size=110084) 2025-01-08T10:47:12,838 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742165_1341 (size=110084) 2025-01-08T10:47:12,881 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742166_1342 (size=1323991) 2025-01-08T10:47:12,882 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742166_1342 (size=1323991) 2025-01-08T10:47:12,883 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742166_1342 (size=1323991) 2025-01-08T10:47:12,916 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742167_1343 (size=23076) 2025-01-08T10:47:12,916 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742167_1343 (size=23076) 2025-01-08T10:47:12,917 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742167_1343 (size=23076) 2025-01-08T10:47:12,941 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742168_1344 (size=126803) 2025-01-08T10:47:12,941 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742168_1344 (size=126803) 2025-01-08T10:47:12,941 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742168_1344 (size=126803) 2025-01-08T10:47:12,956 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742169_1345 (size=322274) 2025-01-08T10:47:12,956 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742169_1345 (size=322274) 2025-01-08T10:47:12,956 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742169_1345 (size=322274) 2025-01-08T10:47:12,983 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742170_1346 (size=1832290) 2025-01-08T10:47:12,984 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742170_1346 (size=1832290) 2025-01-08T10:47:12,984 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742170_1346 (size=1832290) 2025-01-08T10:47:12,997 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742171_1347 (size=30081) 2025-01-08T10:47:12,997 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742171_1347 (size=30081) 2025-01-08T10:47:12,998 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742171_1347 (size=30081) 2025-01-08T10:47:13,006 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742172_1348 (size=53616) 2025-01-08T10:47:13,006 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742172_1348 (size=53616) 2025-01-08T10:47:13,014 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742172_1348 (size=53616) 2025-01-08T10:47:13,016 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742173_1349 (size=29229) 2025-01-08T10:47:13,016 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742173_1349 (size=29229) 2025-01-08T10:47:13,016 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742173_1349 (size=29229) 2025-01-08T10:47:13,024 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742174_1350 (size=169089) 2025-01-08T10:47:13,024 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742174_1350 (size=169089) 2025-01-08T10:47:13,024 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742174_1350 (size=169089) 2025-01-08T10:47:13,047 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742175_1351 (size=6350714) 2025-01-08T10:47:13,047 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742175_1351 (size=6350714) 2025-01-08T10:47:13,051 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742175_1351 (size=6350714) 2025-01-08T10:47:13,080 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742176_1352 (size=5175431) 2025-01-08T10:47:13,081 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742176_1352 (size=5175431) 2025-01-08T10:47:13,081 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742176_1352 (size=5175431) 2025-01-08T10:47:13,089 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742177_1353 (size=136454) 2025-01-08T10:47:13,089 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742177_1353 (size=136454) 2025-01-08T10:47:13,090 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742177_1353 (size=136454) 2025-01-08T10:47:13,100 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742178_1354 (size=912094) 2025-01-08T10:47:13,101 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742178_1354 (size=912094) 2025-01-08T10:47:13,101 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742178_1354 (size=912094) 2025-01-08T10:47:13,118 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742179_1355 (size=3317408) 2025-01-08T10:47:13,119 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742179_1355 (size=3317408) 2025-01-08T10:47:13,119 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742179_1355 (size=3317408) 2025-01-08T10:47:13,164 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742180_1356 (size=451756) 2025-01-08T10:47:13,165 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742180_1356 (size=451756) 2025-01-08T10:47:13,166 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742180_1356 (size=451756) 2025-01-08T10:47:13,201 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742181_1357 (size=503880) 2025-01-08T10:47:13,201 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742181_1357 (size=503880) 2025-01-08T10:47:13,202 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742181_1357 (size=503880) 2025-01-08T10:47:13,240 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742182_1358 (size=4695811) 2025-01-08T10:47:13,240 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742182_1358 (size=4695811) 2025-01-08T10:47:13,241 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742182_1358 (size=4695811) 2025-01-08T10:47:13,243 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T10:47:13,246 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'snaptb0-testExportFileSystemStateWithMergeRegion-1' hfile list 2025-01-08T10:47:13,248 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(751): export split=0 size=9.7 K 2025-01-08T10:47:13,279 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742183_1359 (size=378) 2025-01-08T10:47:13,280 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742183_1359 (size=378) 2025-01-08T10:47:13,281 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742183_1359 (size=378) 2025-01-08T10:47:13,327 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742184_1360 (size=15) 2025-01-08T10:47:13,328 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742184_1360 (size=15) 2025-01-08T10:47:13,328 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742184_1360 (size=15) 2025-01-08T10:47:13,356 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742185_1361 (size=304988) 2025-01-08T10:47:13,357 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742185_1361 (size=304988) 2025-01-08T10:47:13,357 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742185_1361 (size=304988) 2025-01-08T10:47:13,385 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T10:47:13,385 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T10:47:13,596 DEBUG [master/67f42ee69b4e:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 4150f0b85bebbb7094b7a1d91b0a6f32 changed from -1.0 to 0.0, refreshing cache 2025-01-08T10:47:13,597 DEBUG [master/67f42ee69b4e:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 6ff7afa9ac49884297a386018fda40ef changed from -1.0 to 0.0, refreshing cache 2025-01-08T10:47:13,951 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0006_000001 (auth:SIMPLE) from 127.0.0.1:55604 2025-01-08T10:47:17,595 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:17,595 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemStateWithMergeRegion-1 Metrics about Tables on a single HBase RegionServer 2025-01-08T10:47:20,934 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0006_000001 (auth:SIMPLE) from 127.0.0.1:37400 2025-01-08T10:47:21,389 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742186_1362 (size=350662) 2025-01-08T10:47:21,391 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742186_1362 (size=350662) 2025-01-08T10:47:21,391 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742186_1362 (size=350662) 2025-01-08T10:47:23,259 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0006_000001 (auth:SIMPLE) from 127.0.0.1:59034 2025-01-08T10:47:27,342 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742187_1363 (size=4945) 2025-01-08T10:47:27,342 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742187_1363 (size=4945) 2025-01-08T10:47:27,343 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742187_1363 (size=4945) 2025-01-08T10:47:27,398 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742188_1364 (size=4945) 2025-01-08T10:47:27,398 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742188_1364 (size=4945) 2025-01-08T10:47:27,398 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742188_1364 (size=4945) 2025-01-08T10:47:27,482 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742189_1365 (size=17474) 2025-01-08T10:47:27,483 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742189_1365 (size=17474) 2025-01-08T10:47:27,483 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742189_1365 (size=17474) 2025-01-08T10:47:27,497 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742190_1366 (size=482) 2025-01-08T10:47:27,498 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742190_1366 (size=482) 2025-01-08T10:47:27,498 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742190_1366 (size=482) 2025-01-08T10:47:27,552 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742191_1367 (size=17474) 2025-01-08T10:47:27,552 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742191_1367 (size=17474) 2025-01-08T10:47:27,552 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742191_1367 (size=17474) 2025-01-08T10:47:27,559 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_3/usercache/jenkins/appcache/application_1736333116683_0006/container_1736333116683_0006_01_000002/launch_container.sh] 2025-01-08T10:47:27,559 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_3/usercache/jenkins/appcache/application_1736333116683_0006/container_1736333116683_0006_01_000002/container_tokens] 2025-01-08T10:47:27,560 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_3/usercache/jenkins/appcache/application_1736333116683_0006/container_1736333116683_0006_01_000002/sysfs] 2025-01-08T10:47:27,579 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742192_1368 (size=350662) 2025-01-08T10:47:27,579 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742192_1368 (size=350662) 2025-01-08T10:47:27,580 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742192_1368 (size=350662) 2025-01-08T10:47:27,593 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0006_000001 (auth:SIMPLE) from 127.0.0.1:46326 2025-01-08T10:47:29,583 INFO [Time-limited test {}] snapshot.ExportSnapshot(1207): Finalize the Snapshot Export 2025-01-08T10:47:29,585 INFO [Time-limited test {}] snapshot.ExportSnapshot(1218): Verify the exported snapshot's expiration status and integrity. 2025-01-08T10:47:29,593 INFO [Time-limited test {}] snapshot.ExportSnapshot(1224): Export Completed: snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:29,593 INFO [Time-limited test {}] snapshot.TestExportSnapshot(362): Exported snapshot 2025-01-08T10:47:29,593 INFO [Time-limited test {}] snapshot.TestExportSnapshot(373): Verified filesystem state 2025-01-08T10:47:29,593 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-832679797_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1 at hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:29,594 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1/.snapshotinfo 2025-01-08T10:47:29,594 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1/data.manifest 2025-01-08T10:47:29,594 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-832679797_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333230029/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1 at hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333230029/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:29,595 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333230029/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1/.snapshotinfo 2025-01-08T10:47:29,595 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333230029/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1/data.manifest 2025-01-08T10:47:29,602 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:29,603 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.2 disable testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:29,603 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=124, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:29,605 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=124 2025-01-08T10:47:29,605 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion-1","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333249605"}]},"ts":"1736333249605"} 2025-01-08T10:47:29,607 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithMergeRegion-1, state=DISABLING in hbase:meta 2025-01-08T10:47:29,609 INFO [PEWorker-4 {}] procedure.DisableTableProcedure(284): Set testtb-testExportFileSystemStateWithMergeRegion-1 to state=DISABLING 2025-01-08T10:47:29,609 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=125, ppid=124, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1}] 2025-01-08T10:47:29,611 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=126, ppid=125, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=96d4bb39d5de2c0741f205ac02281c75, UNASSIGN}] 2025-01-08T10:47:29,612 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=126, ppid=125, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=96d4bb39d5de2c0741f205ac02281c75, UNASSIGN 2025-01-08T10:47:29,612 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=126 updating hbase:meta row=96d4bb39d5de2c0741f205ac02281c75, regionState=CLOSING, regionLocation=67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:29,614 DEBUG [PEWorker-5 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T10:47:29,614 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=127, ppid=126, state=RUNNABLE; CloseRegionProcedure 96d4bb39d5de2c0741f205ac02281c75, server=67f42ee69b4e,38323,1736333108259}] 2025-01-08T10:47:29,706 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=124 2025-01-08T10:47:29,765 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:29,766 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] handler.UnassignRegionHandler(124): Close 96d4bb39d5de2c0741f205ac02281c75 2025-01-08T10:47:29,767 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T10:47:29,767 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] regionserver.HRegion(1681): Closing 96d4bb39d5de2c0741f205ac02281c75, disabling compactions & flushes 2025-01-08T10:47:29,767 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227962.96d4bb39d5de2c0741f205ac02281c75. 2025-01-08T10:47:29,767 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227962.96d4bb39d5de2c0741f205ac02281c75. 2025-01-08T10:47:29,767 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227962.96d4bb39d5de2c0741f205ac02281c75. after waiting 0 ms 2025-01-08T10:47:29,767 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227962.96d4bb39d5de2c0741f205ac02281c75. 2025-01-08T10:47:29,772 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/96d4bb39d5de2c0741f205ac02281c75/recovered.edits/12.seqid, newMaxSeqId=12, maxSeqId=8 2025-01-08T10:47:29,773 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:47:29,773 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227962.96d4bb39d5de2c0741f205ac02281c75. 2025-01-08T10:47:29,773 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] regionserver.HRegion(1635): Region close journal for 96d4bb39d5de2c0741f205ac02281c75: 2025-01-08T10:47:29,774 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] handler.UnassignRegionHandler(170): Closed 96d4bb39d5de2c0741f205ac02281c75 2025-01-08T10:47:29,775 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=126 updating hbase:meta row=96d4bb39d5de2c0741f205ac02281c75, regionState=CLOSED 2025-01-08T10:47:29,777 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=127, resume processing ppid=126 2025-01-08T10:47:29,777 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=127, ppid=126, state=SUCCESS; CloseRegionProcedure 96d4bb39d5de2c0741f205ac02281c75, server=67f42ee69b4e,38323,1736333108259 in 162 msec 2025-01-08T10:47:29,779 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=126, resume processing ppid=125 2025-01-08T10:47:29,779 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=126, ppid=125, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=96d4bb39d5de2c0741f205ac02281c75, UNASSIGN in 166 msec 2025-01-08T10:47:29,781 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=125, resume processing ppid=124 2025-01-08T10:47:29,781 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=125, ppid=124, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 in 170 msec 2025-01-08T10:47:29,782 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion-1","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333249782"}]},"ts":"1736333249782"} 2025-01-08T10:47:29,783 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithMergeRegion-1, state=DISABLED in hbase:meta 2025-01-08T10:47:29,785 INFO [PEWorker-2 {}] procedure.DisableTableProcedure(296): Set testtb-testExportFileSystemStateWithMergeRegion-1 to state=DISABLED 2025-01-08T10:47:29,787 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=124, state=SUCCESS; DisableTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 in 183 msec 2025-01-08T10:47:29,908 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=124 2025-01-08T10:47:29,908 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion-1, procId: 124 completed 2025-01-08T10:47:29,909 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.2 delete testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:29,909 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=128, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:29,910 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=128, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:29,910 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:29,911 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=128, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:29,912 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36283 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:29,914 DEBUG [HFileArchiver-11 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/96d4bb39d5de2c0741f205ac02281c75 2025-01-08T10:47:29,914 DEBUG [HFileArchiver-12 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/d6f6f260571a44cef14a139fbdf37112 2025-01-08T10:47:29,914 DEBUG [HFileArchiver-13 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/9eb0e33eafa3af0c0eeaba516cde4e33 2025-01-08T10:47:29,915 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:29,915 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:29,916 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:29,916 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:29,916 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion-1 with data PBUF 2025-01-08T10:47:29,916 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion-1 with data PBUF 2025-01-08T10:47:29,916 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion-1 with data PBUF 2025-01-08T10:47:29,916 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion-1 with data PBUF 2025-01-08T10:47:29,917 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:29,917 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:29,917 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:29,918 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:29,918 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:29,918 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:29,918 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:29,918 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:29,919 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=128 2025-01-08T10:47:29,919 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:29,920 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:29,920 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:29,920 DEBUG [HFileArchiver-12 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/d6f6f260571a44cef14a139fbdf37112/cf, FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/d6f6f260571a44cef14a139fbdf37112/recovered.edits] 2025-01-08T10:47:29,920 DEBUG [HFileArchiver-13 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/9eb0e33eafa3af0c0eeaba516cde4e33/cf, FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/9eb0e33eafa3af0c0eeaba516cde4e33/recovered.edits] 2025-01-08T10:47:29,920 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:29,921 DEBUG [HFileArchiver-11 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/96d4bb39d5de2c0741f205ac02281c75/cf, FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/96d4bb39d5de2c0741f205ac02281c75/recovered.edits] 2025-01-08T10:47:29,924 DEBUG [HFileArchiver-13 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/9eb0e33eafa3af0c0eeaba516cde4e33/cf/fe11a77ddadb4b4898e3d882f5f4d1bd to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/9eb0e33eafa3af0c0eeaba516cde4e33/cf/fe11a77ddadb4b4898e3d882f5f4d1bd 2025-01-08T10:47:29,925 DEBUG [HFileArchiver-12 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/d6f6f260571a44cef14a139fbdf37112/cf/a5e18dbe9aa04fce8b1fb88298e33f9a to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/d6f6f260571a44cef14a139fbdf37112/cf/a5e18dbe9aa04fce8b1fb88298e33f9a 2025-01-08T10:47:29,926 DEBUG [HFileArchiver-11 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/96d4bb39d5de2c0741f205ac02281c75/cf/a5e18dbe9aa04fce8b1fb88298e33f9a.d6f6f260571a44cef14a139fbdf37112 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/96d4bb39d5de2c0741f205ac02281c75/cf/a5e18dbe9aa04fce8b1fb88298e33f9a.d6f6f260571a44cef14a139fbdf37112 2025-01-08T10:47:29,927 DEBUG [HFileArchiver-11 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/96d4bb39d5de2c0741f205ac02281c75/cf/fe11a77ddadb4b4898e3d882f5f4d1bd.9eb0e33eafa3af0c0eeaba516cde4e33 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/96d4bb39d5de2c0741f205ac02281c75/cf/fe11a77ddadb4b4898e3d882f5f4d1bd.9eb0e33eafa3af0c0eeaba516cde4e33 2025-01-08T10:47:29,928 DEBUG [HFileArchiver-12 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/d6f6f260571a44cef14a139fbdf37112/recovered.edits/8.seqid to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/d6f6f260571a44cef14a139fbdf37112/recovered.edits/8.seqid 2025-01-08T10:47:29,931 DEBUG [HFileArchiver-13 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/9eb0e33eafa3af0c0eeaba516cde4e33/recovered.edits/8.seqid to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/9eb0e33eafa3af0c0eeaba516cde4e33/recovered.edits/8.seqid 2025-01-08T10:47:29,931 DEBUG [HFileArchiver-12 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/d6f6f260571a44cef14a139fbdf37112 2025-01-08T10:47:29,931 DEBUG [HFileArchiver-13 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/9eb0e33eafa3af0c0eeaba516cde4e33 2025-01-08T10:47:29,933 DEBUG [HFileArchiver-11 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/96d4bb39d5de2c0741f205ac02281c75/recovered.edits/12.seqid to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/96d4bb39d5de2c0741f205ac02281c75/recovered.edits/12.seqid 2025-01-08T10:47:29,933 DEBUG [HFileArchiver-11 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/96d4bb39d5de2c0741f205ac02281c75 2025-01-08T10:47:29,934 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(313): Archived testtb-testExportFileSystemStateWithMergeRegion-1 regions 2025-01-08T10:47:29,936 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=128, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:29,938 WARN [PEWorker-5 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 1 rows of testtb-testExportFileSystemStateWithMergeRegion-1 from hbase:meta 2025-01-08T10:47:29,947 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testExportFileSystemStateWithMergeRegion-1' descriptor. 2025-01-08T10:47:29,949 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=128, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:29,949 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testExportFileSystemStateWithMergeRegion-1' from region states. 2025-01-08T10:47:29,949 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227962.96d4bb39d5de2c0741f205ac02281c75.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736333249949"}]},"ts":"9223372036854775807"} 2025-01-08T10:47:29,953 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1808): Deleted 1 regions from META 2025-01-08T10:47:29,953 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => 96d4bb39d5de2c0741f205ac02281c75, NAME => 'testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227962.96d4bb39d5de2c0741f205ac02281c75.', STARTKEY => '', ENDKEY => ''}] 2025-01-08T10:47:29,953 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testExportFileSystemStateWithMergeRegion-1' as deleted. 2025-01-08T10:47:29,953 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion-1","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736333249953"}]},"ts":"9223372036854775807"} 2025-01-08T10:47:29,956 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testExportFileSystemStateWithMergeRegion-1 state from META 2025-01-08T10:47:29,958 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(133): Finished pid=128, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:29,959 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=128, state=SUCCESS; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 in 50 msec 2025-01-08T10:47:30,021 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=128 2025-01-08T10:47:30,021 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion-1, procId: 128 completed 2025-01-08T10:47:30,021 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:30,022 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.2 disable testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:30,023 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=129, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:30,024 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=129 2025-01-08T10:47:30,024 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333250024"}]},"ts":"1736333250024"} 2025-01-08T10:47:30,026 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithMergeRegion, state=DISABLING in hbase:meta 2025-01-08T10:47:30,033 INFO [PEWorker-1 {}] procedure.DisableTableProcedure(284): Set testtb-testExportFileSystemStateWithMergeRegion to state=DISABLING 2025-01-08T10:47:30,034 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=130, ppid=129, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testExportFileSystemStateWithMergeRegion}] 2025-01-08T10:47:30,035 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=131, ppid=130, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=4150f0b85bebbb7094b7a1d91b0a6f32, UNASSIGN}, {pid=132, ppid=130, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=6ff7afa9ac49884297a386018fda40ef, UNASSIGN}] 2025-01-08T10:47:30,035 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=132, ppid=130, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=6ff7afa9ac49884297a386018fda40ef, UNASSIGN 2025-01-08T10:47:30,036 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=131, ppid=130, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=4150f0b85bebbb7094b7a1d91b0a6f32, UNASSIGN 2025-01-08T10:47:30,036 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=132 updating hbase:meta row=6ff7afa9ac49884297a386018fda40ef, regionState=CLOSING, regionLocation=67f42ee69b4e,45003,1736333108039 2025-01-08T10:47:30,036 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=131 updating hbase:meta row=4150f0b85bebbb7094b7a1d91b0a6f32, regionState=CLOSING, regionLocation=67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:30,038 DEBUG [PEWorker-3 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T10:47:30,038 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=133, ppid=132, state=RUNNABLE; CloseRegionProcedure 6ff7afa9ac49884297a386018fda40ef, server=67f42ee69b4e,45003,1736333108039}] 2025-01-08T10:47:30,038 DEBUG [PEWorker-2 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T10:47:30,038 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=134, ppid=131, state=RUNNABLE; CloseRegionProcedure 4150f0b85bebbb7094b7a1d91b0a6f32, server=67f42ee69b4e,38323,1736333108259}] 2025-01-08T10:47:30,125 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=129 2025-01-08T10:47:30,189 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:47:30,190 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] handler.UnassignRegionHandler(124): Close 6ff7afa9ac49884297a386018fda40ef 2025-01-08T10:47:30,190 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T10:47:30,190 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] regionserver.HRegion(1681): Closing 6ff7afa9ac49884297a386018fda40ef, disabling compactions & flushes 2025-01-08T10:47:30,190 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef. 2025-01-08T10:47:30,190 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef. 2025-01-08T10:47:30,190 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef. after waiting 0 ms 2025-01-08T10:47:30,190 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef. 2025-01-08T10:47:30,190 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:30,191 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] handler.UnassignRegionHandler(124): Close 4150f0b85bebbb7094b7a1d91b0a6f32 2025-01-08T10:47:30,191 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T10:47:30,191 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] regionserver.HRegion(1681): Closing 4150f0b85bebbb7094b7a1d91b0a6f32, disabling compactions & flushes 2025-01-08T10:47:30,191 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32. 2025-01-08T10:47:30,191 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32. 2025-01-08T10:47:30,191 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32. after waiting 0 ms 2025-01-08T10:47:30,191 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32. 2025-01-08T10:47:30,195 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/4150f0b85bebbb7094b7a1d91b0a6f32/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T10:47:30,196 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/6ff7afa9ac49884297a386018fda40ef/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T10:47:30,196 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:47:30,196 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:47:30,196 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32. 2025-01-08T10:47:30,196 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] regionserver.HRegion(1635): Region close journal for 4150f0b85bebbb7094b7a1d91b0a6f32: 2025-01-08T10:47:30,196 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef. 2025-01-08T10:47:30,196 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] regionserver.HRegion(1635): Region close journal for 6ff7afa9ac49884297a386018fda40ef: 2025-01-08T10:47:30,198 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] handler.UnassignRegionHandler(170): Closed 6ff7afa9ac49884297a386018fda40ef 2025-01-08T10:47:30,198 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=132 updating hbase:meta row=6ff7afa9ac49884297a386018fda40ef, regionState=CLOSED 2025-01-08T10:47:30,198 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] handler.UnassignRegionHandler(170): Closed 4150f0b85bebbb7094b7a1d91b0a6f32 2025-01-08T10:47:30,199 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=131 updating hbase:meta row=4150f0b85bebbb7094b7a1d91b0a6f32, regionState=CLOSED 2025-01-08T10:47:30,201 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=133, resume processing ppid=132 2025-01-08T10:47:30,201 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=133, ppid=132, state=SUCCESS; CloseRegionProcedure 6ff7afa9ac49884297a386018fda40ef, server=67f42ee69b4e,45003,1736333108039 in 162 msec 2025-01-08T10:47:30,202 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=134, resume processing ppid=131 2025-01-08T10:47:30,202 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=132, ppid=130, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=6ff7afa9ac49884297a386018fda40ef, UNASSIGN in 166 msec 2025-01-08T10:47:30,203 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=134, ppid=131, state=SUCCESS; CloseRegionProcedure 4150f0b85bebbb7094b7a1d91b0a6f32, server=67f42ee69b4e,38323,1736333108259 in 162 msec 2025-01-08T10:47:30,203 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=131, resume processing ppid=130 2025-01-08T10:47:30,203 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=131, ppid=130, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=4150f0b85bebbb7094b7a1d91b0a6f32, UNASSIGN in 167 msec 2025-01-08T10:47:30,205 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=130, resume processing ppid=129 2025-01-08T10:47:30,205 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=130, ppid=129, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testExportFileSystemStateWithMergeRegion in 171 msec 2025-01-08T10:47:30,206 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333250206"}]},"ts":"1736333250206"} 2025-01-08T10:47:30,207 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithMergeRegion, state=DISABLED in hbase:meta 2025-01-08T10:47:30,208 INFO [PEWorker-4 {}] procedure.DisableTableProcedure(296): Set testtb-testExportFileSystemStateWithMergeRegion to state=DISABLED 2025-01-08T10:47:30,210 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=129, state=SUCCESS; DisableTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion in 187 msec 2025-01-08T10:47:30,326 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=129 2025-01-08T10:47:30,327 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion, procId: 129 completed 2025-01-08T10:47:30,328 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.2 delete testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:30,329 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=135, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:30,330 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=135, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:30,330 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:30,332 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=135, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:30,333 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36283 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:30,337 DEBUG [HFileArchiver-14 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/4150f0b85bebbb7094b7a1d91b0a6f32 2025-01-08T10:47:30,337 DEBUG [HFileArchiver-15 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/6ff7afa9ac49884297a386018fda40ef 2025-01-08T10:47:30,338 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:30,339 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:30,339 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:30,339 DEBUG [HFileArchiver-14 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/4150f0b85bebbb7094b7a1d91b0a6f32/cf, FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/4150f0b85bebbb7094b7a1d91b0a6f32/recovered.edits] 2025-01-08T10:47:30,340 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:30,340 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF 2025-01-08T10:47:30,340 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF 2025-01-08T10:47:30,340 DEBUG [HFileArchiver-15 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/6ff7afa9ac49884297a386018fda40ef/cf, FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/6ff7afa9ac49884297a386018fda40ef/recovered.edits] 2025-01-08T10:47:30,340 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF 2025-01-08T10:47:30,342 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF 2025-01-08T10:47:30,342 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:30,342 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:30,342 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:30,342 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:30,342 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:30,342 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:30,342 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:30,343 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:30,343 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=135 2025-01-08T10:47:30,345 DEBUG [HFileArchiver-14 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/4150f0b85bebbb7094b7a1d91b0a6f32/cf/b85d87029a5646189c7a994b7c99cdba to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion/4150f0b85bebbb7094b7a1d91b0a6f32/cf/b85d87029a5646189c7a994b7c99cdba 2025-01-08T10:47:30,345 DEBUG [HFileArchiver-15 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/6ff7afa9ac49884297a386018fda40ef/cf/cd90586f90f64d5691980aaf0b4c9136 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion/6ff7afa9ac49884297a386018fda40ef/cf/cd90586f90f64d5691980aaf0b4c9136 2025-01-08T10:47:30,347 DEBUG [HFileArchiver-14 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/4150f0b85bebbb7094b7a1d91b0a6f32/recovered.edits/9.seqid to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion/4150f0b85bebbb7094b7a1d91b0a6f32/recovered.edits/9.seqid 2025-01-08T10:47:30,348 DEBUG [HFileArchiver-14 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/4150f0b85bebbb7094b7a1d91b0a6f32 2025-01-08T10:47:30,349 DEBUG [HFileArchiver-15 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/6ff7afa9ac49884297a386018fda40ef/recovered.edits/9.seqid to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion/6ff7afa9ac49884297a386018fda40ef/recovered.edits/9.seqid 2025-01-08T10:47:30,349 DEBUG [HFileArchiver-15 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithMergeRegion/6ff7afa9ac49884297a386018fda40ef 2025-01-08T10:47:30,349 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(313): Archived testtb-testExportFileSystemStateWithMergeRegion regions 2025-01-08T10:47:30,350 DEBUG [PEWorker-2 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c 2025-01-08T10:47:30,351 DEBUG [PEWorker-2 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf] 2025-01-08T10:47:30,354 DEBUG [PEWorker-2 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/c4ca4238a0b923820dcc509a6f75849b2025010810301e755fe74f01a4220a74bb7a8296_6ff7afa9ac49884297a386018fda40ef to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/c4ca4238a0b923820dcc509a6f75849b2025010810301e755fe74f01a4220a74bb7a8296_6ff7afa9ac49884297a386018fda40ef 2025-01-08T10:47:30,355 DEBUG [PEWorker-2 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/d41d8cd98f00b204e9800998ecf8427e20250108df8d023a67d946d4a1716e125356d5e2_4150f0b85bebbb7094b7a1d91b0a6f32 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/d41d8cd98f00b204e9800998ecf8427e20250108df8d023a67d946d4a1716e125356d5e2_4150f0b85bebbb7094b7a1d91b0a6f32 2025-01-08T10:47:30,356 DEBUG [PEWorker-2 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c 2025-01-08T10:47:30,360 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=135, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:30,362 WARN [PEWorker-2 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testtb-testExportFileSystemStateWithMergeRegion from hbase:meta 2025-01-08T10:47:30,364 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testExportFileSystemStateWithMergeRegion' descriptor. 2025-01-08T10:47:30,365 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=135, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:30,365 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testExportFileSystemStateWithMergeRegion' from region states. 2025-01-08T10:47:30,366 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736333250365"}]},"ts":"9223372036854775807"} 2025-01-08T10:47:30,366 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736333250365"}]},"ts":"9223372036854775807"} 2025-01-08T10:47:30,367 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T10:47:30,367 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => 4150f0b85bebbb7094b7a1d91b0a6f32, NAME => 'testtb-testExportFileSystemStateWithMergeRegion,,1736333225301.4150f0b85bebbb7094b7a1d91b0a6f32.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => 6ff7afa9ac49884297a386018fda40ef, NAME => 'testtb-testExportFileSystemStateWithMergeRegion,1,1736333225301.6ff7afa9ac49884297a386018fda40ef.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T10:47:30,367 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testExportFileSystemStateWithMergeRegion' as deleted. 2025-01-08T10:47:30,367 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736333250367"}]},"ts":"9223372036854775807"} 2025-01-08T10:47:30,369 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testExportFileSystemStateWithMergeRegion state from META 2025-01-08T10:47:30,371 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(133): Finished pid=135, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:30,372 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=135, state=SUCCESS; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion in 43 msec 2025-01-08T10:47:30,444 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=135 2025-01-08T10:47:30,445 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion, procId: 135 completed 2025-01-08T10:47:30,453 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.2 delete name: "emptySnaptb0-testExportFileSystemStateWithMergeRegion" 2025-01-08T10:47:30,454 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(380): Deleting snapshot: emptySnaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:30,456 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.2 delete name: "snaptb0-testExportFileSystemStateWithMergeRegion" 2025-01-08T10:47:30,457 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:30,459 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.2 delete name: "snaptb0-testExportFileSystemStateWithMergeRegion-1" 2025-01-08T10:47:30,461 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:30,485 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: snapshot.TestMobSecureExportSnapshot#testExportFileSystemStateWithMergeRegion Thread=799 (was 787) Potentially hanging thread: HFileArchiver-15 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-6-7 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:62) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:883) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-30 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LogDeleter #3 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1177) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-392475539_1 at /127.0.0.1:45856 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-12 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-13 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_COMPACTED_FILES_DISCHARGER-regionserver/67f42ee69b4e:0-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-34 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-11 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:39653 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-33 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-6-5 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:62) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:883) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-14 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Thread-4956 java.base@17.0.11/java.io.FileInputStream.readBytes(Native Method) java.base@17.0.11/java.io.FileInputStream.read(FileInputStream.java:276) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:282) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) java.base@17.0.11/sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:281) java.base@17.0.11/sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:324) java.base@17.0.11/sun.nio.cs.StreamDecoder.read(StreamDecoder.java:189) java.base@17.0.11/java.io.InputStreamReader.read(InputStreamReader.java:177) java.base@17.0.11/java.io.BufferedReader.fill(BufferedReader.java:162) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:329) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:396) app//org.apache.hadoop.util.Shell$1.run(Shell.java:1025) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-832679797_22 at /127.0.0.1:57992 [Waiting for operation #6] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_COMPACTED_FILES_DISCHARGER-regionserver/67f42ee69b4e:0-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-832679797_22 at /127.0.0.1:45884 [Waiting for operation #4] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_COMPACTED_FILES_DISCHARGER-regionserver/67f42ee69b4e:0-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:37895 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #10 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-6-6 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:62) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:883) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-832679797_22 at /127.0.0.1:39212 [Waiting for operation #4] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #9 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-392475539_1 at /127.0.0.1:57968 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: process reaper (pid 20191) java.base@17.0.11/java.lang.ProcessHandleImpl.waitForProcessExit0(Native Method) java.base@17.0.11/java.lang.ProcessHandleImpl$1.run(ProcessHandleImpl.java:150) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (175901922) connection to localhost/127.0.0.1:39653 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-31 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (175901922) connection to localhost/127.0.0.1:37895 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-32 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-29 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) - Thread LEAK? -, OpenFileDescriptor=815 (was 803) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=875 (was 841) - SystemLoadAverage LEAK? -, ProcessCount=15 (was 18), AvailableMemoryMB=1503 (was 1473) - AvailableMemoryMB LEAK? - 2025-01-08T10:47:30,485 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=799 is superior to 500 2025-01-08T10:47:30,508 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: snapshot.TestMobSecureExportSnapshot#testExportExpiredSnapshot Thread=799, OpenFileDescriptor=815, MaxFileDescriptor=1048576, SystemLoadAverage=875, ProcessCount=16, AvailableMemoryMB=1497 2025-01-08T10:47:30,508 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=799 is superior to 500 2025-01-08T10:47:30,510 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.2 create 'testtb-testExportExpiredSnapshot', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T10:47:30,511 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=136, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testExportExpiredSnapshot 2025-01-08T10:47:30,512 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=136, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportExpiredSnapshot execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T10:47:30,512 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "testtb-testExportExpiredSnapshot" procId is: 136 2025-01-08T10:47:30,513 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=136, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testExportExpiredSnapshot execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T10:47:30,514 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=136 2025-01-08T10:47:30,519 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742193_1369 (size=443) 2025-01-08T10:47:30,520 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742193_1369 (size=443) 2025-01-08T10:47:30,520 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742193_1369 (size=443) 2025-01-08T10:47:30,522 INFO [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 85afd80183d737be3864ee9b52caa712, NAME => 'testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testtb-testExportExpiredSnapshot', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:47:30,522 INFO [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => 97a34fb309b891785b6b66cdc5d5f74b, NAME => 'testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testtb-testExportExpiredSnapshot', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:47:30,536 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742194_1370 (size=68) 2025-01-08T10:47:30,536 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742194_1370 (size=68) 2025-01-08T10:47:30,536 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742194_1370 (size=68) 2025-01-08T10:47:30,537 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:47:30,537 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1681): Closing 85afd80183d737be3864ee9b52caa712, disabling compactions & flushes 2025-01-08T10:47:30,537 INFO [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712. 2025-01-08T10:47:30,537 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712. 2025-01-08T10:47:30,537 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712. after waiting 0 ms 2025-01-08T10:47:30,537 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712. 2025-01-08T10:47:30,537 INFO [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712. 2025-01-08T10:47:30,537 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1635): Region close journal for 85afd80183d737be3864ee9b52caa712: 2025-01-08T10:47:30,544 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742195_1371 (size=68) 2025-01-08T10:47:30,545 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742195_1371 (size=68) 2025-01-08T10:47:30,545 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742195_1371 (size=68) 2025-01-08T10:47:30,545 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:47:30,546 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1681): Closing 97a34fb309b891785b6b66cdc5d5f74b, disabling compactions & flushes 2025-01-08T10:47:30,546 INFO [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b. 2025-01-08T10:47:30,546 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b. 2025-01-08T10:47:30,546 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b. after waiting 0 ms 2025-01-08T10:47:30,546 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b. 2025-01-08T10:47:30,546 INFO [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b. 2025-01-08T10:47:30,546 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1635): Region close journal for 97a34fb309b891785b6b66cdc5d5f74b: 2025-01-08T10:47:30,547 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=136, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testExportExpiredSnapshot execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T10:47:30,548 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712.","families":{"info":[{"qualifier":"regioninfo","vlen":67,"tag":[],"timestamp":"1736333250547"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736333250547"}]},"ts":"1736333250547"} 2025-01-08T10:47:30,548 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b.","families":{"info":[{"qualifier":"regioninfo","vlen":67,"tag":[],"timestamp":"1736333250547"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736333250547"}]},"ts":"1736333250547"} 2025-01-08T10:47:30,550 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T10:47:30,550 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=136, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testExportExpiredSnapshot execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T10:47:30,551 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportExpiredSnapshot","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333250550"}]},"ts":"1736333250550"} 2025-01-08T10:47:30,552 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportExpiredSnapshot, state=ENABLING in hbase:meta 2025-01-08T10:47:30,555 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(202): Hosts are {67f42ee69b4e=0} racks are {/default-rack=0} 2025-01-08T10:47:30,557 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T10:47:30,557 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T10:47:30,557 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T10:47:30,557 INFO [PEWorker-3 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T10:47:30,557 INFO [PEWorker-3 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T10:47:30,557 INFO [PEWorker-3 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T10:47:30,557 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T10:47:30,557 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=137, ppid=136, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=85afd80183d737be3864ee9b52caa712, ASSIGN}, {pid=138, ppid=136, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=97a34fb309b891785b6b66cdc5d5f74b, ASSIGN}] 2025-01-08T10:47:30,558 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=138, ppid=136, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=97a34fb309b891785b6b66cdc5d5f74b, ASSIGN 2025-01-08T10:47:30,558 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=137, ppid=136, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=85afd80183d737be3864ee9b52caa712, ASSIGN 2025-01-08T10:47:30,560 INFO [PEWorker-5 {}] assignment.TransitRegionStateProcedure(264): Starting pid=138, ppid=136, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=97a34fb309b891785b6b66cdc5d5f74b, ASSIGN; state=OFFLINE, location=67f42ee69b4e,45003,1736333108039; forceNewPlan=false, retain=false 2025-01-08T10:47:30,560 INFO [PEWorker-1 {}] assignment.TransitRegionStateProcedure(264): Starting pid=137, ppid=136, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=85afd80183d737be3864ee9b52caa712, ASSIGN; state=OFFLINE, location=67f42ee69b4e,36283,1736333108166; forceNewPlan=false, retain=false 2025-01-08T10:47:30,615 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=136 2025-01-08T10:47:30,710 INFO [67f42ee69b4e:34551 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T10:47:30,711 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=137 updating hbase:meta row=85afd80183d737be3864ee9b52caa712, regionState=OPENING, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:30,711 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=138 updating hbase:meta row=97a34fb309b891785b6b66cdc5d5f74b, regionState=OPENING, regionLocation=67f42ee69b4e,45003,1736333108039 2025-01-08T10:47:30,712 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=139, ppid=138, state=RUNNABLE; OpenRegionProcedure 97a34fb309b891785b6b66cdc5d5f74b, server=67f42ee69b4e,45003,1736333108039}] 2025-01-08T10:47:30,713 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=140, ppid=137, state=RUNNABLE; OpenRegionProcedure 85afd80183d737be3864ee9b52caa712, server=67f42ee69b4e,36283,1736333108166}] 2025-01-08T10:47:30,816 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=136 2025-01-08T10:47:30,864 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:47:30,865 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:30,868 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] handler.AssignRegionHandler(135): Open testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b. 2025-01-08T10:47:30,868 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(7285): Opening region: {ENCODED => 97a34fb309b891785b6b66cdc5d5f74b, NAME => 'testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T10:47:30,868 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b. service=AccessControlService 2025-01-08T10:47:30,869 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] handler.AssignRegionHandler(135): Open testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712. 2025-01-08T10:47:30,869 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:47:30,869 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(7285): Opening region: {ENCODED => 85afd80183d737be3864ee9b52caa712, NAME => 'testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T10:47:30,869 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportExpiredSnapshot 97a34fb309b891785b6b66cdc5d5f74b 2025-01-08T10:47:30,869 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(894): Instantiated testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:47:30,869 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(7327): checking encryption for 97a34fb309b891785b6b66cdc5d5f74b 2025-01-08T10:47:30,869 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(7330): checking classloading for 97a34fb309b891785b6b66cdc5d5f74b 2025-01-08T10:47:30,869 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712. service=AccessControlService 2025-01-08T10:47:30,869 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:47:30,869 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportExpiredSnapshot 85afd80183d737be3864ee9b52caa712 2025-01-08T10:47:30,869 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(894): Instantiated testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:47:30,869 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(7327): checking encryption for 85afd80183d737be3864ee9b52caa712 2025-01-08T10:47:30,869 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(7330): checking classloading for 85afd80183d737be3864ee9b52caa712 2025-01-08T10:47:30,870 INFO [StoreOpener-97a34fb309b891785b6b66cdc5d5f74b-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 97a34fb309b891785b6b66cdc5d5f74b 2025-01-08T10:47:30,872 INFO [StoreOpener-97a34fb309b891785b6b66cdc5d5f74b-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 97a34fb309b891785b6b66cdc5d5f74b columnFamilyName cf 2025-01-08T10:47:30,873 DEBUG [StoreOpener-97a34fb309b891785b6b66cdc5d5f74b-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:30,873 INFO [StoreOpener-97a34fb309b891785b6b66cdc5d5f74b-1 {}] regionserver.HStore(327): Store=97a34fb309b891785b6b66cdc5d5f74b/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:47:30,874 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/97a34fb309b891785b6b66cdc5d5f74b 2025-01-08T10:47:30,874 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/97a34fb309b891785b6b66cdc5d5f74b 2025-01-08T10:47:30,877 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(1085): writing seq id for 97a34fb309b891785b6b66cdc5d5f74b 2025-01-08T10:47:30,886 INFO [StoreOpener-85afd80183d737be3864ee9b52caa712-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 85afd80183d737be3864ee9b52caa712 2025-01-08T10:47:30,888 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/97a34fb309b891785b6b66cdc5d5f74b/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:47:30,888 INFO [StoreOpener-85afd80183d737be3864ee9b52caa712-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 85afd80183d737be3864ee9b52caa712 columnFamilyName cf 2025-01-08T10:47:30,888 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(1102): Opened 97a34fb309b891785b6b66cdc5d5f74b; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=66632693, jitterRate=-0.007095500826835632}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T10:47:30,889 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(1001): Region open journal for 97a34fb309b891785b6b66cdc5d5f74b: 2025-01-08T10:47:30,890 DEBUG [StoreOpener-85afd80183d737be3864ee9b52caa712-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:30,890 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b., pid=139, masterSystemTime=1736333250864 2025-01-08T10:47:30,890 INFO [StoreOpener-85afd80183d737be3864ee9b52caa712-1 {}] regionserver.HStore(327): Store=85afd80183d737be3864ee9b52caa712/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:47:30,891 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/85afd80183d737be3864ee9b52caa712 2025-01-08T10:47:30,891 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b. 2025-01-08T10:47:30,891 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] handler.AssignRegionHandler(164): Opened testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b. 2025-01-08T10:47:30,892 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/85afd80183d737be3864ee9b52caa712 2025-01-08T10:47:30,892 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=138 updating hbase:meta row=97a34fb309b891785b6b66cdc5d5f74b, regionState=OPEN, openSeqNum=2, regionLocation=67f42ee69b4e,45003,1736333108039 2025-01-08T10:47:30,894 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(1085): writing seq id for 85afd80183d737be3864ee9b52caa712 2025-01-08T10:47:30,896 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=139, resume processing ppid=138 2025-01-08T10:47:30,896 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=139, ppid=138, state=SUCCESS; OpenRegionProcedure 97a34fb309b891785b6b66cdc5d5f74b, server=67f42ee69b4e,45003,1736333108039 in 182 msec 2025-01-08T10:47:30,896 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/85afd80183d737be3864ee9b52caa712/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:47:30,897 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(1102): Opened 85afd80183d737be3864ee9b52caa712; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=70740901, jitterRate=0.05412156879901886}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T10:47:30,897 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=138, ppid=136, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=97a34fb309b891785b6b66cdc5d5f74b, ASSIGN in 339 msec 2025-01-08T10:47:30,897 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(1001): Region open journal for 85afd80183d737be3864ee9b52caa712: 2025-01-08T10:47:30,898 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712., pid=140, masterSystemTime=1736333250865 2025-01-08T10:47:30,899 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712. 2025-01-08T10:47:30,899 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] handler.AssignRegionHandler(164): Opened testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712. 2025-01-08T10:47:30,899 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=137 updating hbase:meta row=85afd80183d737be3864ee9b52caa712, regionState=OPEN, openSeqNum=2, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:30,902 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=140, resume processing ppid=137 2025-01-08T10:47:30,902 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=140, ppid=137, state=SUCCESS; OpenRegionProcedure 85afd80183d737be3864ee9b52caa712, server=67f42ee69b4e,36283,1736333108166 in 188 msec 2025-01-08T10:47:30,904 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=137, resume processing ppid=136 2025-01-08T10:47:30,904 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=137, ppid=136, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=85afd80183d737be3864ee9b52caa712, ASSIGN in 345 msec 2025-01-08T10:47:30,905 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=136, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testExportExpiredSnapshot execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T10:47:30,905 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportExpiredSnapshot","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333250905"}]},"ts":"1736333250905"} 2025-01-08T10:47:30,906 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportExpiredSnapshot, state=ENABLED in hbase:meta 2025-01-08T10:47:30,909 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=136, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportExpiredSnapshot execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T10:47:30,909 DEBUG [PEWorker-5 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testExportExpiredSnapshot jenkins: RWXCA 2025-01-08T10:47:30,911 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36283 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportExpiredSnapshot], kv [jenkins: RWXCA] 2025-01-08T10:47:30,913 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:30,913 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:30,913 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:30,913 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:30,915 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportExpiredSnapshot 2025-01-08T10:47:30,915 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportExpiredSnapshot 2025-01-08T10:47:30,915 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportExpiredSnapshot 2025-01-08T10:47:30,916 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportExpiredSnapshot 2025-01-08T10:47:30,916 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:30,916 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:30,916 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:30,916 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:30,917 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=136, state=SUCCESS; CreateTableProcedure table=testtb-testExportExpiredSnapshot in 406 msec 2025-01-08T10:47:31,117 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=136 2025-01-08T10:47:31,117 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testExportExpiredSnapshot, procId: 136 completed 2025-01-08T10:47:31,120 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportExpiredSnapshot 2025-01-08T10:47:31,120 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712. 2025-01-08T10:47:31,120 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T10:47:31,131 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.2 snapshot request for:{ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } 2025-01-08T10:47:31,132 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736333251132 (current time:1736333251132). 2025-01-08T10:47:31,132 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T10:47:31,132 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot emptySnaptb0-testExportExpiredSnapshot VERSION not specified, setting to 2 2025-01-08T10:47:31,132 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T10:47:31,132 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x153428c8 to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@aa55e72 2025-01-08T10:47:31,135 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@37aae881, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:47:31,137 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:47:31,138 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:52800, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:47:31,139 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x153428c8 to 127.0.0.1:52367 2025-01-08T10:47:31,139 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:47:31,140 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x004529a0 to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@2033a114 2025-01-08T10:47:31,143 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@403dbb73, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:47:31,144 DEBUG [hconnection-0x6d3f2951-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:47:31,145 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:52814, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:47:31,148 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x004529a0 to 127.0.0.1:52367 2025-01-08T10:47:31,148 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:47:31,148 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportExpiredSnapshot], kv [jenkins: RWXCA] 2025-01-08T10:47:31,149 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T10:47:31,149 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=141, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } 2025-01-08T10:47:31,150 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 }, snapshot procedure id = 141 2025-01-08T10:47:31,150 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T10:47:31,150 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=141 2025-01-08T10:47:31,151 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T10:47:31,153 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T10:47:31,162 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742196_1372 (size=170) 2025-01-08T10:47:31,162 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742196_1372 (size=170) 2025-01-08T10:47:31,164 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742196_1372 (size=170) 2025-01-08T10:47:31,165 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T10:47:31,166 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=142, ppid=141, state=RUNNABLE; SnapshotRegionProcedure 85afd80183d737be3864ee9b52caa712}, {pid=143, ppid=141, state=RUNNABLE; SnapshotRegionProcedure 97a34fb309b891785b6b66cdc5d5f74b}] 2025-01-08T10:47:31,166 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=143, ppid=141, state=RUNNABLE; SnapshotRegionProcedure 97a34fb309b891785b6b66cdc5d5f74b 2025-01-08T10:47:31,167 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=142, ppid=141, state=RUNNABLE; SnapshotRegionProcedure 85afd80183d737be3864ee9b52caa712 2025-01-08T10:47:31,252 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=141 2025-01-08T10:47:31,317 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:47:31,317 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:31,318 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=45003 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=143 2025-01-08T10:47:31,318 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=36283 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=142 2025-01-08T10:47:31,318 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=142}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712. 2025-01-08T10:47:31,318 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=143}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b. 2025-01-08T10:47:31,318 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=142}] regionserver.HRegion(2538): Flush status journal for 85afd80183d737be3864ee9b52caa712: 2025-01-08T10:47:31,318 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=143}] regionserver.HRegion(2538): Flush status journal for 97a34fb309b891785b6b66cdc5d5f74b: 2025-01-08T10:47:31,318 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=142}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712. for emptySnaptb0-testExportExpiredSnapshot completed. 2025-01-08T10:47:31,318 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=143}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b. for emptySnaptb0-testExportExpiredSnapshot completed. 2025-01-08T10:47:31,319 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=142}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712.' region-info for snapshot=emptySnaptb0-testExportExpiredSnapshot 2025-01-08T10:47:31,319 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=142}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:47:31,319 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=142}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T10:47:31,319 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=143}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b.' region-info for snapshot=emptySnaptb0-testExportExpiredSnapshot 2025-01-08T10:47:31,319 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=143}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:47:31,319 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=143}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T10:47:31,327 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742197_1373 (size=71) 2025-01-08T10:47:31,328 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742197_1373 (size=71) 2025-01-08T10:47:31,328 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742197_1373 (size=71) 2025-01-08T10:47:31,328 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=142}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712. 2025-01-08T10:47:31,328 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=142}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=142 2025-01-08T10:47:31,328 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=142 2025-01-08T10:47:31,329 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportExpiredSnapshot on region 85afd80183d737be3864ee9b52caa712 2025-01-08T10:47:31,329 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=142, ppid=141, state=RUNNABLE; SnapshotRegionProcedure 85afd80183d737be3864ee9b52caa712 2025-01-08T10:47:31,330 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=142, ppid=141, state=SUCCESS; SnapshotRegionProcedure 85afd80183d737be3864ee9b52caa712 in 163 msec 2025-01-08T10:47:31,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742198_1374 (size=71) 2025-01-08T10:47:31,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742198_1374 (size=71) 2025-01-08T10:47:31,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742198_1374 (size=71) 2025-01-08T10:47:31,333 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=143}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b. 2025-01-08T10:47:31,333 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=143}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=143 2025-01-08T10:47:31,333 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=143 2025-01-08T10:47:31,333 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportExpiredSnapshot on region 97a34fb309b891785b6b66cdc5d5f74b 2025-01-08T10:47:31,334 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=143, ppid=141, state=RUNNABLE; SnapshotRegionProcedure 97a34fb309b891785b6b66cdc5d5f74b 2025-01-08T10:47:31,335 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=143, resume processing ppid=141 2025-01-08T10:47:31,335 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=143, ppid=141, state=SUCCESS; SnapshotRegionProcedure 97a34fb309b891785b6b66cdc5d5f74b in 168 msec 2025-01-08T10:47:31,336 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T10:47:31,336 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T10:47:31,337 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T10:47:31,337 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T10:47:31,337 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:31,337 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(328): No files under family: cf 2025-01-08T10:47:31,343 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742199_1375 (size=63) 2025-01-08T10:47:31,343 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742199_1375 (size=63) 2025-01-08T10:47:31,344 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742199_1375 (size=63) 2025-01-08T10:47:31,345 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T10:47:31,345 DEBUG [PEWorker-4 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for emptySnaptb0-testExportExpiredSnapshot 2025-01-08T10:47:31,345 DEBUG [PEWorker-4 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/emptySnaptb0-testExportExpiredSnapshot 2025-01-08T10:47:31,362 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742200_1376 (size=653) 2025-01-08T10:47:31,362 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742200_1376 (size=653) 2025-01-08T10:47:31,362 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742200_1376 (size=653) 2025-01-08T10:47:31,370 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T10:47:31,375 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T10:47:31,375 DEBUG [PEWorker-4 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/emptySnaptb0-testExportExpiredSnapshot to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/emptySnaptb0-testExportExpiredSnapshot 2025-01-08T10:47:31,377 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T10:47:31,377 DEBUG [PEWorker-4 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 }, snapshot procedure id = 141 2025-01-08T10:47:31,378 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=141, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } in 228 msec 2025-01-08T10:47:31,453 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=141 2025-01-08T10:47:31,453 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportExpiredSnapshot, procId: 141 completed 2025-01-08T10:47:31,455 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36283 {}] regionserver.HRegion(8254): writing data to region testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T10:47:31,456 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=45003 {}] regionserver.HRegion(8254): writing data to region testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T10:47:31,458 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportExpiredSnapshot 2025-01-08T10:47:31,459 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712. 2025-01-08T10:47:31,459 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T10:47:31,469 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.2 snapshot request for:{ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } 2025-01-08T10:47:31,469 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736333251469 (current time:1736333251469). 2025-01-08T10:47:31,469 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T10:47:31,469 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testExportExpiredSnapshot VERSION not specified, setting to 2 2025-01-08T10:47:31,469 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T10:47:31,470 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x54d3fd9a to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@403223f0 2025-01-08T10:47:31,476 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@74b37374, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:47:31,478 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:47:31,479 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:52826, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:47:31,480 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x54d3fd9a to 127.0.0.1:52367 2025-01-08T10:47:31,480 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:47:31,482 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x07d82ec6 to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@56a4a06e 2025-01-08T10:47:31,487 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@39e3511d, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:47:31,489 DEBUG [hconnection-0x144dd050-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:47:31,490 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:52838, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:47:31,492 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x07d82ec6 to 127.0.0.1:52367 2025-01-08T10:47:31,492 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:47:31,492 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportExpiredSnapshot], kv [jenkins: RWXCA] 2025-01-08T10:47:31,493 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T10:47:31,494 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=144, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } 2025-01-08T10:47:31,494 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 }, snapshot procedure id = 144 2025-01-08T10:47:31,495 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T10:47:31,495 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=144 2025-01-08T10:47:31,495 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T10:47:31,497 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T10:47:31,504 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742201_1377 (size=165) 2025-01-08T10:47:31,505 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742201_1377 (size=165) 2025-01-08T10:47:31,505 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742201_1377 (size=165) 2025-01-08T10:47:31,506 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T10:47:31,506 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=145, ppid=144, state=RUNNABLE; SnapshotRegionProcedure 85afd80183d737be3864ee9b52caa712}, {pid=146, ppid=144, state=RUNNABLE; SnapshotRegionProcedure 97a34fb309b891785b6b66cdc5d5f74b}] 2025-01-08T10:47:31,507 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=146, ppid=144, state=RUNNABLE; SnapshotRegionProcedure 97a34fb309b891785b6b66cdc5d5f74b 2025-01-08T10:47:31,507 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=145, ppid=144, state=RUNNABLE; SnapshotRegionProcedure 85afd80183d737be3864ee9b52caa712 2025-01-08T10:47:31,596 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=144 2025-01-08T10:47:31,658 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:47:31,658 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:31,659 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=36283 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=145 2025-01-08T10:47:31,659 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=45003 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=146 2025-01-08T10:47:31,659 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b. 2025-01-08T10:47:31,659 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712. 2025-01-08T10:47:31,659 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.HRegion(2837): Flushing 85afd80183d737be3864ee9b52caa712 1/1 column families, dataSize=199 B heapSize=688 B 2025-01-08T10:47:31,659 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.HRegion(2837): Flushing 97a34fb309b891785b6b66cdc5d5f74b 1/1 column families, dataSize=3.06 KB heapSize=6.86 KB 2025-01-08T10:47:31,685 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e202501085524df7972ac45c8a7e12cf1e21547b4_85afd80183d737be3864ee9b52caa712 is 71, key is 000b1389d693ca5771e18dc92b3afc1c/cf:q/1736333251455/Put/seqid=0 2025-01-08T10:47:31,685 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108dcdf324001a04f5ba9431dc262de604d_97a34fb309b891785b6b66cdc5d5f74b is 71, key is 10453512fbdd82dd13103e8de8f0accb/cf:q/1736333251456/Put/seqid=0 2025-01-08T10:47:31,704 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742203_1379 (size=5102) 2025-01-08T10:47:31,704 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742203_1379 (size=5102) 2025-01-08T10:47:31,704 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742203_1379 (size=5102) 2025-01-08T10:47:31,705 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:31,709 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e202501085524df7972ac45c8a7e12cf1e21547b4_85afd80183d737be3864ee9b52caa712 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/d41d8cd98f00b204e9800998ecf8427e202501085524df7972ac45c8a7e12cf1e21547b4_85afd80183d737be3864ee9b52caa712 2025-01-08T10:47:31,710 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/85afd80183d737be3864ee9b52caa712/.tmp/cf/63eee5a987bc4a5ab241a37776a04aed, store: [table=testtb-testExportExpiredSnapshot family=cf region=85afd80183d737be3864ee9b52caa712] 2025-01-08T10:47:31,711 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/85afd80183d737be3864ee9b52caa712/.tmp/cf/63eee5a987bc4a5ab241a37776a04aed is 209, key is 027a33cb353af64cd0b761d9bb3855b12/cf:q/1736333251455/Put/seqid=0 2025-01-08T10:47:31,719 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742202_1378 (size=8171) 2025-01-08T10:47:31,719 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742202_1378 (size=8171) 2025-01-08T10:47:31,720 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742202_1378 (size=8171) 2025-01-08T10:47:31,721 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:31,732 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108dcdf324001a04f5ba9431dc262de604d_97a34fb309b891785b6b66cdc5d5f74b to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/c4ca4238a0b923820dcc509a6f75849b20250108dcdf324001a04f5ba9431dc262de604d_97a34fb309b891785b6b66cdc5d5f74b 2025-01-08T10:47:31,733 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/97a34fb309b891785b6b66cdc5d5f74b/.tmp/cf/6a68e14d37d14e47a8305cff873d55fe, store: [table=testtb-testExportExpiredSnapshot family=cf region=97a34fb309b891785b6b66cdc5d5f74b] 2025-01-08T10:47:31,734 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/97a34fb309b891785b6b66cdc5d5f74b/.tmp/cf/6a68e14d37d14e47a8305cff873d55fe is 209, key is 1df2e455efdafbc6b6e609857431cfa9d/cf:q/1736333251456/Put/seqid=0 2025-01-08T10:47:31,746 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742204_1380 (size=5918) 2025-01-08T10:47:31,747 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742204_1380 (size=5918) 2025-01-08T10:47:31,747 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742204_1380 (size=5918) 2025-01-08T10:47:31,747 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=199, hasBloomFilter=true, into tmp file hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/85afd80183d737be3864ee9b52caa712/.tmp/cf/63eee5a987bc4a5ab241a37776a04aed 2025-01-08T10:47:31,753 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/85afd80183d737be3864ee9b52caa712/.tmp/cf/63eee5a987bc4a5ab241a37776a04aed as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/85afd80183d737be3864ee9b52caa712/cf/63eee5a987bc4a5ab241a37776a04aed 2025-01-08T10:47:31,757 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/85afd80183d737be3864ee9b52caa712/cf/63eee5a987bc4a5ab241a37776a04aed, entries=3, sequenceid=6, filesize=5.8 K 2025-01-08T10:47:31,758 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.HRegion(3040): Finished flush of dataSize ~199 B/199, heapSize ~672 B/672, currentSize=0 B/0 for 85afd80183d737be3864ee9b52caa712 in 99ms, sequenceid=6, compaction requested=false 2025-01-08T10:47:31,758 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportExpiredSnapshot' 2025-01-08T10:47:31,759 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.HRegion(2538): Flush status journal for 85afd80183d737be3864ee9b52caa712: 2025-01-08T10:47:31,759 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712. for snaptb0-testExportExpiredSnapshot completed. 2025-01-08T10:47:31,759 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712.' region-info for snapshot=snaptb0-testExportExpiredSnapshot 2025-01-08T10:47:31,759 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:47:31,759 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/85afd80183d737be3864ee9b52caa712/cf/63eee5a987bc4a5ab241a37776a04aed] hfiles 2025-01-08T10:47:31,759 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/85afd80183d737be3864ee9b52caa712/cf/63eee5a987bc4a5ab241a37776a04aed for snapshot=snaptb0-testExportExpiredSnapshot 2025-01-08T10:47:31,764 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742205_1381 (size=14997) 2025-01-08T10:47:31,764 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742205_1381 (size=14997) 2025-01-08T10:47:31,765 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742205_1381 (size=14997) 2025-01-08T10:47:31,765 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=3.1 K, hasBloomFilter=true, into tmp file hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/97a34fb309b891785b6b66cdc5d5f74b/.tmp/cf/6a68e14d37d14e47a8305cff873d55fe 2025-01-08T10:47:31,771 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/97a34fb309b891785b6b66cdc5d5f74b/.tmp/cf/6a68e14d37d14e47a8305cff873d55fe as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/97a34fb309b891785b6b66cdc5d5f74b/cf/6a68e14d37d14e47a8305cff873d55fe 2025-01-08T10:47:31,776 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/97a34fb309b891785b6b66cdc5d5f74b/cf/6a68e14d37d14e47a8305cff873d55fe, entries=47, sequenceid=6, filesize=14.6 K 2025-01-08T10:47:31,777 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.HRegion(3040): Finished flush of dataSize ~3.06 KB/3137, heapSize ~6.84 KB/7008, currentSize=0 B/0 for 97a34fb309b891785b6b66cdc5d5f74b in 118ms, sequenceid=6, compaction requested=false 2025-01-08T10:47:31,777 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.HRegion(2538): Flush status journal for 97a34fb309b891785b6b66cdc5d5f74b: 2025-01-08T10:47:31,777 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b. for snaptb0-testExportExpiredSnapshot completed. 2025-01-08T10:47:31,777 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b.' region-info for snapshot=snaptb0-testExportExpiredSnapshot 2025-01-08T10:47:31,777 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:47:31,777 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/97a34fb309b891785b6b66cdc5d5f74b/cf/6a68e14d37d14e47a8305cff873d55fe] hfiles 2025-01-08T10:47:31,777 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/97a34fb309b891785b6b66cdc5d5f74b/cf/6a68e14d37d14e47a8305cff873d55fe for snapshot=snaptb0-testExportExpiredSnapshot 2025-01-08T10:47:31,797 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=144 2025-01-08T10:47:31,827 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742206_1382 (size=110) 2025-01-08T10:47:31,828 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742206_1382 (size=110) 2025-01-08T10:47:31,828 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742206_1382 (size=110) 2025-01-08T10:47:31,828 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712. 2025-01-08T10:47:31,828 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=145 2025-01-08T10:47:31,829 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=145 2025-01-08T10:47:31,829 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportExpiredSnapshot on region 85afd80183d737be3864ee9b52caa712 2025-01-08T10:47:31,829 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=145, ppid=144, state=RUNNABLE; SnapshotRegionProcedure 85afd80183d737be3864ee9b52caa712 2025-01-08T10:47:31,831 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=145, ppid=144, state=SUCCESS; SnapshotRegionProcedure 85afd80183d737be3864ee9b52caa712 in 324 msec 2025-01-08T10:47:31,872 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742207_1383 (size=110) 2025-01-08T10:47:31,873 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742207_1383 (size=110) 2025-01-08T10:47:31,873 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742207_1383 (size=110) 2025-01-08T10:47:31,874 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b. 2025-01-08T10:47:31,874 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=146 2025-01-08T10:47:31,874 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=146 2025-01-08T10:47:31,874 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportExpiredSnapshot on region 97a34fb309b891785b6b66cdc5d5f74b 2025-01-08T10:47:31,875 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=146, ppid=144, state=RUNNABLE; SnapshotRegionProcedure 97a34fb309b891785b6b66cdc5d5f74b 2025-01-08T10:47:31,878 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=146, resume processing ppid=144 2025-01-08T10:47:31,879 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T10:47:31,879 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=146, ppid=144, state=SUCCESS; SnapshotRegionProcedure 97a34fb309b891785b6b66cdc5d5f74b in 370 msec 2025-01-08T10:47:31,880 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T10:47:31,891 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T10:47:31,891 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T10:47:31,892 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:31,893 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/c4ca4238a0b923820dcc509a6f75849b20250108dcdf324001a04f5ba9431dc262de604d_97a34fb309b891785b6b66cdc5d5f74b, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/d41d8cd98f00b204e9800998ecf8427e202501085524df7972ac45c8a7e12cf1e21547b4_85afd80183d737be3864ee9b52caa712] hfiles 2025-01-08T10:47:31,893 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/c4ca4238a0b923820dcc509a6f75849b20250108dcdf324001a04f5ba9431dc262de604d_97a34fb309b891785b6b66cdc5d5f74b 2025-01-08T10:47:31,894 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/d41d8cd98f00b204e9800998ecf8427e202501085524df7972ac45c8a7e12cf1e21547b4_85afd80183d737be3864ee9b52caa712 2025-01-08T10:47:31,950 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742208_1384 (size=294) 2025-01-08T10:47:31,950 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742208_1384 (size=294) 2025-01-08T10:47:31,950 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742208_1384 (size=294) 2025-01-08T10:47:32,099 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=144 2025-01-08T10:47:32,351 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T10:47:32,351 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testExportExpiredSnapshot 2025-01-08T10:47:32,352 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/snaptb0-testExportExpiredSnapshot 2025-01-08T10:47:32,378 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742209_1385 (size=963) 2025-01-08T10:47:32,379 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742209_1385 (size=963) 2025-01-08T10:47:32,379 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742209_1385 (size=963) 2025-01-08T10:47:32,385 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T10:47:32,405 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T10:47:32,406 DEBUG [PEWorker-3 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/snaptb0-testExportExpiredSnapshot to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportExpiredSnapshot 2025-01-08T10:47:32,407 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T10:47:32,407 DEBUG [PEWorker-3 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 }, snapshot procedure id = 144 2025-01-08T10:47:32,409 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=144, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } in 914 msec 2025-01-08T10:47:32,600 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=144 2025-01-08T10:47:32,600 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportExpiredSnapshot, procId: 144 completed 2025-01-08T10:47:32,602 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.2 create 'testExportExpiredSnapshot', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T10:47:32,602 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=147, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testExportExpiredSnapshot 2025-01-08T10:47:32,603 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=147, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testExportExpiredSnapshot execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T10:47:32,603 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "testExportExpiredSnapshot" procId is: 147 2025-01-08T10:47:32,604 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=147, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testExportExpiredSnapshot execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T10:47:32,604 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=147 2025-01-08T10:47:32,617 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742210_1386 (size=436) 2025-01-08T10:47:32,620 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742210_1386 (size=436) 2025-01-08T10:47:32,620 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742210_1386 (size=436) 2025-01-08T10:47:32,622 INFO [RegionOpenAndInit-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => a8e8f2d8192bce47f7fedf056a7a63d9, NAME => 'testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testExportExpiredSnapshot', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:47:32,623 INFO [RegionOpenAndInit-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => 5f4e73fdd950eac2bfc23f9cc1241d47, NAME => 'testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testExportExpiredSnapshot', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:47:32,652 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742211_1387 (size=61) 2025-01-08T10:47:32,652 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742211_1387 (size=61) 2025-01-08T10:47:32,667 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742211_1387 (size=61) 2025-01-08T10:47:32,669 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(894): Instantiated testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:47:32,669 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1681): Closing a8e8f2d8192bce47f7fedf056a7a63d9, disabling compactions & flushes 2025-01-08T10:47:32,669 INFO [RegionOpenAndInit-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1703): Closing region testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9. 2025-01-08T10:47:32,669 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9. 2025-01-08T10:47:32,669 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9. after waiting 0 ms 2025-01-08T10:47:32,669 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9. 2025-01-08T10:47:32,669 INFO [RegionOpenAndInit-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1922): Closed testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9. 2025-01-08T10:47:32,669 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1635): Region close journal for a8e8f2d8192bce47f7fedf056a7a63d9: 2025-01-08T10:47:32,672 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742212_1388 (size=61) 2025-01-08T10:47:32,673 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742212_1388 (size=61) 2025-01-08T10:47:32,674 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742212_1388 (size=61) 2025-01-08T10:47:32,675 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(894): Instantiated testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:47:32,675 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1681): Closing 5f4e73fdd950eac2bfc23f9cc1241d47, disabling compactions & flushes 2025-01-08T10:47:32,675 INFO [RegionOpenAndInit-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1703): Closing region testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47. 2025-01-08T10:47:32,675 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47. 2025-01-08T10:47:32,675 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47. after waiting 0 ms 2025-01-08T10:47:32,675 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47. 2025-01-08T10:47:32,675 INFO [RegionOpenAndInit-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1922): Closed testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47. 2025-01-08T10:47:32,675 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1635): Region close journal for 5f4e73fdd950eac2bfc23f9cc1241d47: 2025-01-08T10:47:32,677 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=147, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testExportExpiredSnapshot execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T10:47:32,677 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9.","families":{"info":[{"qualifier":"regioninfo","vlen":60,"tag":[],"timestamp":"1736333252677"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736333252677"}]},"ts":"1736333252677"} 2025-01-08T10:47:32,677 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47.","families":{"info":[{"qualifier":"regioninfo","vlen":60,"tag":[],"timestamp":"1736333252677"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736333252677"}]},"ts":"1736333252677"} 2025-01-08T10:47:32,681 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T10:47:32,682 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=147, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testExportExpiredSnapshot execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T10:47:32,683 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testExportExpiredSnapshot","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333252682"}]},"ts":"1736333252682"} 2025-01-08T10:47:32,687 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1655): Updated tableName=testExportExpiredSnapshot, state=ENABLING in hbase:meta 2025-01-08T10:47:32,692 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(202): Hosts are {67f42ee69b4e=0} racks are {/default-rack=0} 2025-01-08T10:47:32,694 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T10:47:32,694 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T10:47:32,694 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T10:47:32,694 INFO [PEWorker-1 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T10:47:32,694 INFO [PEWorker-1 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T10:47:32,694 INFO [PEWorker-1 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T10:47:32,694 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T10:47:32,694 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=148, ppid=147, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testExportExpiredSnapshot, region=a8e8f2d8192bce47f7fedf056a7a63d9, ASSIGN}, {pid=149, ppid=147, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testExportExpiredSnapshot, region=5f4e73fdd950eac2bfc23f9cc1241d47, ASSIGN}] 2025-01-08T10:47:32,696 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=149, ppid=147, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testExportExpiredSnapshot, region=5f4e73fdd950eac2bfc23f9cc1241d47, ASSIGN 2025-01-08T10:47:32,696 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=148, ppid=147, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testExportExpiredSnapshot, region=a8e8f2d8192bce47f7fedf056a7a63d9, ASSIGN 2025-01-08T10:47:32,698 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=149, ppid=147, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testExportExpiredSnapshot, region=5f4e73fdd950eac2bfc23f9cc1241d47, ASSIGN; state=OFFLINE, location=67f42ee69b4e,45003,1736333108039; forceNewPlan=false, retain=false 2025-01-08T10:47:32,698 INFO [PEWorker-5 {}] assignment.TransitRegionStateProcedure(264): Starting pid=148, ppid=147, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testExportExpiredSnapshot, region=a8e8f2d8192bce47f7fedf056a7a63d9, ASSIGN; state=OFFLINE, location=67f42ee69b4e,36283,1736333108166; forceNewPlan=false, retain=false 2025-01-08T10:47:32,705 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=147 2025-01-08T10:47:32,848 INFO [67f42ee69b4e:34551 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T10:47:32,849 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=148 updating hbase:meta row=a8e8f2d8192bce47f7fedf056a7a63d9, regionState=OPENING, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:32,849 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=149 updating hbase:meta row=5f4e73fdd950eac2bfc23f9cc1241d47, regionState=OPENING, regionLocation=67f42ee69b4e,45003,1736333108039 2025-01-08T10:47:32,852 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=150, ppid=148, state=RUNNABLE; OpenRegionProcedure a8e8f2d8192bce47f7fedf056a7a63d9, server=67f42ee69b4e,36283,1736333108166}] 2025-01-08T10:47:32,853 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=151, ppid=149, state=RUNNABLE; OpenRegionProcedure 5f4e73fdd950eac2bfc23f9cc1241d47, server=67f42ee69b4e,45003,1736333108039}] 2025-01-08T10:47:32,907 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=147 2025-01-08T10:47:33,005 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:33,005 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:47:33,009 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] handler.AssignRegionHandler(135): Open testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47. 2025-01-08T10:47:33,009 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(7285): Opening region: {ENCODED => 5f4e73fdd950eac2bfc23f9cc1241d47, NAME => 'testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T10:47:33,009 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] handler.AssignRegionHandler(135): Open testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9. 2025-01-08T10:47:33,009 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(7999): Registered coprocessor service: region=testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47. service=AccessControlService 2025-01-08T10:47:33,009 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(7285): Opening region: {ENCODED => a8e8f2d8192bce47f7fedf056a7a63d9, NAME => 'testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T10:47:33,009 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:47:33,010 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(7999): Registered coprocessor service: region=testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9. service=AccessControlService 2025-01-08T10:47:33,010 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testExportExpiredSnapshot 5f4e73fdd950eac2bfc23f9cc1241d47 2025-01-08T10:47:33,010 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(894): Instantiated testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:47:33,010 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:47:33,010 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(7327): checking encryption for 5f4e73fdd950eac2bfc23f9cc1241d47 2025-01-08T10:47:33,010 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(7330): checking classloading for 5f4e73fdd950eac2bfc23f9cc1241d47 2025-01-08T10:47:33,010 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testExportExpiredSnapshot a8e8f2d8192bce47f7fedf056a7a63d9 2025-01-08T10:47:33,010 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(894): Instantiated testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:47:33,010 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(7327): checking encryption for a8e8f2d8192bce47f7fedf056a7a63d9 2025-01-08T10:47:33,010 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(7330): checking classloading for a8e8f2d8192bce47f7fedf056a7a63d9 2025-01-08T10:47:33,012 INFO [StoreOpener-a8e8f2d8192bce47f7fedf056a7a63d9-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region a8e8f2d8192bce47f7fedf056a7a63d9 2025-01-08T10:47:33,012 INFO [StoreOpener-5f4e73fdd950eac2bfc23f9cc1241d47-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 5f4e73fdd950eac2bfc23f9cc1241d47 2025-01-08T10:47:33,014 INFO [StoreOpener-a8e8f2d8192bce47f7fedf056a7a63d9-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region a8e8f2d8192bce47f7fedf056a7a63d9 columnFamilyName cf 2025-01-08T10:47:33,014 INFO [StoreOpener-5f4e73fdd950eac2bfc23f9cc1241d47-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 5f4e73fdd950eac2bfc23f9cc1241d47 columnFamilyName cf 2025-01-08T10:47:33,015 DEBUG [StoreOpener-a8e8f2d8192bce47f7fedf056a7a63d9-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:33,015 DEBUG [StoreOpener-5f4e73fdd950eac2bfc23f9cc1241d47-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:33,016 INFO [StoreOpener-a8e8f2d8192bce47f7fedf056a7a63d9-1 {}] regionserver.HStore(327): Store=a8e8f2d8192bce47f7fedf056a7a63d9/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:47:33,016 INFO [StoreOpener-5f4e73fdd950eac2bfc23f9cc1241d47-1 {}] regionserver.HStore(327): Store=5f4e73fdd950eac2bfc23f9cc1241d47/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:47:33,017 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportExpiredSnapshot/a8e8f2d8192bce47f7fedf056a7a63d9 2025-01-08T10:47:33,017 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportExpiredSnapshot/5f4e73fdd950eac2bfc23f9cc1241d47 2025-01-08T10:47:33,017 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportExpiredSnapshot/a8e8f2d8192bce47f7fedf056a7a63d9 2025-01-08T10:47:33,017 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportExpiredSnapshot/5f4e73fdd950eac2bfc23f9cc1241d47 2025-01-08T10:47:33,019 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(1085): writing seq id for a8e8f2d8192bce47f7fedf056a7a63d9 2025-01-08T10:47:33,019 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(1085): writing seq id for 5f4e73fdd950eac2bfc23f9cc1241d47 2025-01-08T10:47:33,021 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportExpiredSnapshot/5f4e73fdd950eac2bfc23f9cc1241d47/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:47:33,021 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportExpiredSnapshot/a8e8f2d8192bce47f7fedf056a7a63d9/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:47:33,022 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(1102): Opened a8e8f2d8192bce47f7fedf056a7a63d9; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=65046202, jitterRate=-0.030736058950424194}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T10:47:33,022 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(1102): Opened 5f4e73fdd950eac2bfc23f9cc1241d47; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=64419331, jitterRate=-0.04007716476917267}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T10:47:33,022 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(1001): Region open journal for a8e8f2d8192bce47f7fedf056a7a63d9: 2025-01-08T10:47:33,022 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(1001): Region open journal for 5f4e73fdd950eac2bfc23f9cc1241d47: 2025-01-08T10:47:33,023 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegionServer(2601): Post open deploy tasks for testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9., pid=150, masterSystemTime=1736333253005 2025-01-08T10:47:33,023 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegionServer(2601): Post open deploy tasks for testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47., pid=151, masterSystemTime=1736333253005 2025-01-08T10:47:33,025 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegionServer(2628): Finished post open deploy task for testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9. 2025-01-08T10:47:33,025 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] handler.AssignRegionHandler(164): Opened testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9. 2025-01-08T10:47:33,025 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=148 updating hbase:meta row=a8e8f2d8192bce47f7fedf056a7a63d9, regionState=OPEN, openSeqNum=2, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:33,026 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegionServer(2628): Finished post open deploy task for testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47. 2025-01-08T10:47:33,026 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] handler.AssignRegionHandler(164): Opened testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47. 2025-01-08T10:47:33,027 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=149 updating hbase:meta row=5f4e73fdd950eac2bfc23f9cc1241d47, regionState=OPEN, openSeqNum=2, regionLocation=67f42ee69b4e,45003,1736333108039 2025-01-08T10:47:33,038 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=150, resume processing ppid=148 2025-01-08T10:47:33,038 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=150, ppid=148, state=SUCCESS; OpenRegionProcedure a8e8f2d8192bce47f7fedf056a7a63d9, server=67f42ee69b4e,36283,1736333108166 in 176 msec 2025-01-08T10:47:33,042 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=151, resume processing ppid=149 2025-01-08T10:47:33,042 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=151, ppid=149, state=SUCCESS; OpenRegionProcedure 5f4e73fdd950eac2bfc23f9cc1241d47, server=67f42ee69b4e,45003,1736333108039 in 176 msec 2025-01-08T10:47:33,043 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=148, ppid=147, state=SUCCESS; TransitRegionStateProcedure table=testExportExpiredSnapshot, region=a8e8f2d8192bce47f7fedf056a7a63d9, ASSIGN in 344 msec 2025-01-08T10:47:33,048 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=149, resume processing ppid=147 2025-01-08T10:47:33,048 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=149, ppid=147, state=SUCCESS; TransitRegionStateProcedure table=testExportExpiredSnapshot, region=5f4e73fdd950eac2bfc23f9cc1241d47, ASSIGN in 348 msec 2025-01-08T10:47:33,050 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=147, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testExportExpiredSnapshot execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T10:47:33,051 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testExportExpiredSnapshot","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333253051"}]},"ts":"1736333253051"} 2025-01-08T10:47:33,053 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testExportExpiredSnapshot, state=ENABLED in hbase:meta 2025-01-08T10:47:33,058 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=147, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testExportExpiredSnapshot execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T10:47:33,058 DEBUG [PEWorker-2 {}] access.PermissionStorage(175): Writing permission with rowKey testExportExpiredSnapshot jenkins: RWXCA 2025-01-08T10:47:33,062 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36283 {}] access.PermissionStorage(611): Read acl: entry[testExportExpiredSnapshot], kv [jenkins: RWXCA] 2025-01-08T10:47:33,064 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:33,064 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:33,064 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:33,064 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:33,066 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:33,067 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:33,067 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:33,067 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:33,068 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:33,068 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:33,068 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:33,068 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:33,069 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=147, state=SUCCESS; CreateTableProcedure table=testExportExpiredSnapshot in 464 msec 2025-01-08T10:47:33,208 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=147 2025-01-08T10:47:33,208 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testExportExpiredSnapshot, procId: 147 completed 2025-01-08T10:47:33,210 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testExportExpiredSnapshot 2025-01-08T10:47:33,210 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9. 2025-01-08T10:47:33,211 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T10:47:33,227 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36283 {}] regionserver.HRegion(8254): writing data to region testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T10:47:33,228 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=45003 {}] regionserver.HRegion(8254): writing data to region testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T10:47:33,231 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testExportExpiredSnapshot 2025-01-08T10:47:33,231 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9. 2025-01-08T10:47:33,231 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T10:47:33,240 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.2 snapshot request for:{ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } 2025-01-08T10:47:33,240 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snapshot-testExportExpiredSnapshot VERSION not specified, setting to 2 2025-01-08T10:47:33,240 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T10:47:33,241 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x0140965c to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@2867df58 2025-01-08T10:47:33,247 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@64d979e4, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:47:33,248 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:47:33,249 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:52854, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:47:33,250 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x0140965c to 127.0.0.1:52367 2025-01-08T10:47:33,250 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:47:33,251 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x7b89a54a to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@6d02387c 2025-01-08T10:47:33,254 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@423b6a61, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:47:33,256 DEBUG [hconnection-0x4d3f2f4f-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:47:33,257 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:52858, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:47:33,258 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x7b89a54a to 127.0.0.1:52367 2025-01-08T10:47:33,258 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:47:33,258 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] access.PermissionStorage(611): Read acl: entry[testExportExpiredSnapshot], kv [jenkins: RWXCA] 2025-01-08T10:47:33,259 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T10:47:33,260 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=152, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } 2025-01-08T10:47:33,260 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 }, snapshot procedure id = 152 2025-01-08T10:47:33,261 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_PREPARE 2025-01-08T10:47:33,261 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=152 2025-01-08T10:47:33,261 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T10:47:33,264 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T10:47:33,272 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742213_1389 (size=152) 2025-01-08T10:47:33,272 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742213_1389 (size=152) 2025-01-08T10:47:33,272 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742213_1389 (size=152) 2025-01-08T10:47:33,273 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T10:47:33,273 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=153, ppid=152, state=RUNNABLE; SnapshotRegionProcedure a8e8f2d8192bce47f7fedf056a7a63d9}, {pid=154, ppid=152, state=RUNNABLE; SnapshotRegionProcedure 5f4e73fdd950eac2bfc23f9cc1241d47}] 2025-01-08T10:47:33,274 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=154, ppid=152, state=RUNNABLE; SnapshotRegionProcedure 5f4e73fdd950eac2bfc23f9cc1241d47 2025-01-08T10:47:33,274 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=153, ppid=152, state=RUNNABLE; SnapshotRegionProcedure a8e8f2d8192bce47f7fedf056a7a63d9 2025-01-08T10:47:33,362 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=152 2025-01-08T10:47:33,425 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:47:33,425 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:33,426 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=45003 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=154 2025-01-08T10:47:33,426 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=36283 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=153 2025-01-08T10:47:33,426 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9. 2025-01-08T10:47:33,426 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47. 2025-01-08T10:47:33,426 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.HRegion(2837): Flushing a8e8f2d8192bce47f7fedf056a7a63d9 1/1 column families, dataSize=132 B heapSize=544 B 2025-01-08T10:47:33,426 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.HRegion(2837): Flushing 5f4e73fdd950eac2bfc23f9cc1241d47 1/1 column families, dataSize=3.13 KB heapSize=7 KB 2025-01-08T10:47:33,452 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e202501083b181fc615774c88902b47fe8390b4b8_a8e8f2d8192bce47f7fedf056a7a63d9 is 71, key is 00812475012f005b6a78dd667590f682/cf:q/1736333253227/Put/seqid=0 2025-01-08T10:47:33,456 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108805228b6b6964a38a941651ee4f5bdfb_5f4e73fdd950eac2bfc23f9cc1241d47 is 71, key is 15ddb011ebebc44e93572e1e18f9c94a/cf:q/1736333253228/Put/seqid=0 2025-01-08T10:47:33,467 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742214_1390 (size=5032) 2025-01-08T10:47:33,468 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742214_1390 (size=5032) 2025-01-08T10:47:33,468 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742214_1390 (size=5032) 2025-01-08T10:47:33,469 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:33,476 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e202501083b181fc615774c88902b47fe8390b4b8_a8e8f2d8192bce47f7fedf056a7a63d9 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testExportExpiredSnapshot/1030f41967fbb659ab4c2a7a1774d313/cf/d41d8cd98f00b204e9800998ecf8427e202501083b181fc615774c88902b47fe8390b4b8_a8e8f2d8192bce47f7fedf056a7a63d9 2025-01-08T10:47:33,477 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportExpiredSnapshot/a8e8f2d8192bce47f7fedf056a7a63d9/.tmp/cf/c11126505cd04f39a3ab0f013aa2d42e, store: [table=testExportExpiredSnapshot family=cf region=a8e8f2d8192bce47f7fedf056a7a63d9] 2025-01-08T10:47:33,478 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportExpiredSnapshot/a8e8f2d8192bce47f7fedf056a7a63d9/.tmp/cf/c11126505cd04f39a3ab0f013aa2d42e is 202, key is 04a91bf2fb06ef79f552034322e759dc9/cf:q/1736333253227/Put/seqid=0 2025-01-08T10:47:33,484 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742215_1391 (size=8241) 2025-01-08T10:47:33,485 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742215_1391 (size=8241) 2025-01-08T10:47:33,485 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742215_1391 (size=8241) 2025-01-08T10:47:33,485 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:33,491 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108805228b6b6964a38a941651ee4f5bdfb_5f4e73fdd950eac2bfc23f9cc1241d47 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testExportExpiredSnapshot/1030f41967fbb659ab4c2a7a1774d313/cf/c4ca4238a0b923820dcc509a6f75849b20250108805228b6b6964a38a941651ee4f5bdfb_5f4e73fdd950eac2bfc23f9cc1241d47 2025-01-08T10:47:33,492 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportExpiredSnapshot/5f4e73fdd950eac2bfc23f9cc1241d47/.tmp/cf/47bfe478f41f4b97b52f80cc6049df3d, store: [table=testExportExpiredSnapshot family=cf region=5f4e73fdd950eac2bfc23f9cc1241d47] 2025-01-08T10:47:33,493 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportExpiredSnapshot/5f4e73fdd950eac2bfc23f9cc1241d47/.tmp/cf/47bfe478f41f4b97b52f80cc6049df3d is 202, key is 16e94ea5778796114a2e91e88c6e614e4/cf:q/1736333253228/Put/seqid=0 2025-01-08T10:47:33,513 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742217_1393 (size=14859) 2025-01-08T10:47:33,514 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742217_1393 (size=14859) 2025-01-08T10:47:33,514 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742217_1393 (size=14859) 2025-01-08T10:47:33,515 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=5, memsize=3.1 K, hasBloomFilter=true, into tmp file hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportExpiredSnapshot/5f4e73fdd950eac2bfc23f9cc1241d47/.tmp/cf/47bfe478f41f4b97b52f80cc6049df3d 2025-01-08T10:47:33,525 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportExpiredSnapshot/5f4e73fdd950eac2bfc23f9cc1241d47/.tmp/cf/47bfe478f41f4b97b52f80cc6049df3d as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportExpiredSnapshot/5f4e73fdd950eac2bfc23f9cc1241d47/cf/47bfe478f41f4b97b52f80cc6049df3d 2025-01-08T10:47:33,529 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742216_1392 (size=5688) 2025-01-08T10:47:33,529 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742216_1392 (size=5688) 2025-01-08T10:47:33,530 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742216_1392 (size=5688) 2025-01-08T10:47:33,531 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=5, memsize=132, hasBloomFilter=true, into tmp file hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportExpiredSnapshot/a8e8f2d8192bce47f7fedf056a7a63d9/.tmp/cf/c11126505cd04f39a3ab0f013aa2d42e 2025-01-08T10:47:33,531 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportExpiredSnapshot/5f4e73fdd950eac2bfc23f9cc1241d47/cf/47bfe478f41f4b97b52f80cc6049df3d, entries=48, sequenceid=5, filesize=14.5 K 2025-01-08T10:47:33,534 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.HRegion(3040): Finished flush of dataSize ~3.13 KB/3204, heapSize ~6.98 KB/7152, currentSize=0 B/0 for 5f4e73fdd950eac2bfc23f9cc1241d47 in 108ms, sequenceid=5, compaction requested=false 2025-01-08T10:47:33,534 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testExportExpiredSnapshot' 2025-01-08T10:47:33,535 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.HRegion(2538): Flush status journal for 5f4e73fdd950eac2bfc23f9cc1241d47: 2025-01-08T10:47:33,535 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.SnapshotRegionCallable(75): Snapshotting region testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47. for snapshot-testExportExpiredSnapshot completed. 2025-01-08T10:47:33,535 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] snapshot.SnapshotManifest(241): Storing 'testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47.' region-info for snapshot=snapshot-testExportExpiredSnapshot 2025-01-08T10:47:33,535 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:47:33,535 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportExpiredSnapshot/5f4e73fdd950eac2bfc23f9cc1241d47/cf/47bfe478f41f4b97b52f80cc6049df3d] hfiles 2025-01-08T10:47:33,535 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportExpiredSnapshot/5f4e73fdd950eac2bfc23f9cc1241d47/cf/47bfe478f41f4b97b52f80cc6049df3d for snapshot=snapshot-testExportExpiredSnapshot 2025-01-08T10:47:33,541 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportExpiredSnapshot/a8e8f2d8192bce47f7fedf056a7a63d9/.tmp/cf/c11126505cd04f39a3ab0f013aa2d42e as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportExpiredSnapshot/a8e8f2d8192bce47f7fedf056a7a63d9/cf/c11126505cd04f39a3ab0f013aa2d42e 2025-01-08T10:47:33,547 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportExpiredSnapshot/a8e8f2d8192bce47f7fedf056a7a63d9/cf/c11126505cd04f39a3ab0f013aa2d42e, entries=2, sequenceid=5, filesize=5.6 K 2025-01-08T10:47:33,548 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.HRegion(3040): Finished flush of dataSize ~132 B/132, heapSize ~528 B/528, currentSize=0 B/0 for a8e8f2d8192bce47f7fedf056a7a63d9 in 122ms, sequenceid=5, compaction requested=false 2025-01-08T10:47:33,548 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.HRegion(2538): Flush status journal for a8e8f2d8192bce47f7fedf056a7a63d9: 2025-01-08T10:47:33,548 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.SnapshotRegionCallable(75): Snapshotting region testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9. for snapshot-testExportExpiredSnapshot completed. 2025-01-08T10:47:33,548 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] snapshot.SnapshotManifest(241): Storing 'testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9.' region-info for snapshot=snapshot-testExportExpiredSnapshot 2025-01-08T10:47:33,548 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:47:33,548 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportExpiredSnapshot/a8e8f2d8192bce47f7fedf056a7a63d9/cf/c11126505cd04f39a3ab0f013aa2d42e] hfiles 2025-01-08T10:47:33,548 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportExpiredSnapshot/a8e8f2d8192bce47f7fedf056a7a63d9/cf/c11126505cd04f39a3ab0f013aa2d42e for snapshot=snapshot-testExportExpiredSnapshot 2025-01-08T10:47:33,564 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=152 2025-01-08T10:47:33,601 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742218_1394 (size=103) 2025-01-08T10:47:33,602 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742218_1394 (size=103) 2025-01-08T10:47:33,602 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742218_1394 (size=103) 2025-01-08T10:47:33,603 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47. 2025-01-08T10:47:33,603 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=154 2025-01-08T10:47:33,603 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=154 2025-01-08T10:47:33,603 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snapshot-testExportExpiredSnapshot on region 5f4e73fdd950eac2bfc23f9cc1241d47 2025-01-08T10:47:33,604 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=154, ppid=152, state=RUNNABLE; SnapshotRegionProcedure 5f4e73fdd950eac2bfc23f9cc1241d47 2025-01-08T10:47:33,605 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742219_1395 (size=103) 2025-01-08T10:47:33,606 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742219_1395 (size=103) 2025-01-08T10:47:33,606 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=154, ppid=152, state=SUCCESS; SnapshotRegionProcedure 5f4e73fdd950eac2bfc23f9cc1241d47 in 332 msec 2025-01-08T10:47:33,607 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742219_1395 (size=103) 2025-01-08T10:47:33,608 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9. 2025-01-08T10:47:33,608 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=153 2025-01-08T10:47:33,608 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=153 2025-01-08T10:47:33,608 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snapshot-testExportExpiredSnapshot on region a8e8f2d8192bce47f7fedf056a7a63d9 2025-01-08T10:47:33,609 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=153, ppid=152, state=RUNNABLE; SnapshotRegionProcedure a8e8f2d8192bce47f7fedf056a7a63d9 2025-01-08T10:47:33,612 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=153, resume processing ppid=152 2025-01-08T10:47:33,612 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=153, ppid=152, state=SUCCESS; SnapshotRegionProcedure a8e8f2d8192bce47f7fedf056a7a63d9 in 338 msec 2025-01-08T10:47:33,612 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T10:47:33,613 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T10:47:33,615 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T10:47:33,615 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T10:47:33,615 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:33,616 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testExportExpiredSnapshot/1030f41967fbb659ab4c2a7a1774d313/cf/c4ca4238a0b923820dcc509a6f75849b20250108805228b6b6964a38a941651ee4f5bdfb_5f4e73fdd950eac2bfc23f9cc1241d47, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testExportExpiredSnapshot/1030f41967fbb659ab4c2a7a1774d313/cf/d41d8cd98f00b204e9800998ecf8427e202501083b181fc615774c88902b47fe8390b4b8_a8e8f2d8192bce47f7fedf056a7a63d9] hfiles 2025-01-08T10:47:33,616 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testExportExpiredSnapshot/1030f41967fbb659ab4c2a7a1774d313/cf/c4ca4238a0b923820dcc509a6f75849b20250108805228b6b6964a38a941651ee4f5bdfb_5f4e73fdd950eac2bfc23f9cc1241d47 2025-01-08T10:47:33,616 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testExportExpiredSnapshot/1030f41967fbb659ab4c2a7a1774d313/cf/d41d8cd98f00b204e9800998ecf8427e202501083b181fc615774c88902b47fe8390b4b8_a8e8f2d8192bce47f7fedf056a7a63d9 2025-01-08T10:47:33,640 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742220_1396 (size=287) 2025-01-08T10:47:33,640 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742220_1396 (size=287) 2025-01-08T10:47:33,641 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742220_1396 (size=287) 2025-01-08T10:47:33,642 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T10:47:33,642 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snapshot-testExportExpiredSnapshot 2025-01-08T10:47:33,643 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/snapshot-testExportExpiredSnapshot 2025-01-08T10:47:33,684 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742221_1397 (size=935) 2025-01-08T10:47:33,684 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742221_1397 (size=935) 2025-01-08T10:47:33,684 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742221_1397 (size=935) 2025-01-08T10:47:33,687 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T10:47:33,692 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T10:47:33,693 DEBUG [PEWorker-1 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/snapshot-testExportExpiredSnapshot to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snapshot-testExportExpiredSnapshot 2025-01-08T10:47:33,695 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T10:47:33,695 DEBUG [PEWorker-1 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 }, snapshot procedure id = 152 2025-01-08T10:47:33,698 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=152, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } in 436 msec 2025-01-08T10:47:33,704 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0006_000001 (auth:SIMPLE) from 127.0.0.1:52756 2025-01-08T10:47:33,715 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_2/usercache/jenkins/appcache/application_1736333116683_0006/container_1736333116683_0006_01_000001/launch_container.sh] 2025-01-08T10:47:33,715 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_2/usercache/jenkins/appcache/application_1736333116683_0006/container_1736333116683_0006_01_000001/container_tokens] 2025-01-08T10:47:33,715 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_2/usercache/jenkins/appcache/application_1736333116683_0006/container_1736333116683_0006_01_000001/sysfs] 2025-01-08T10:47:33,865 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=152 2025-01-08T10:47:33,865 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testExportExpiredSnapshot, procId: 152 completed 2025-01-08T10:47:35,308 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T10:47:35,714 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T10:47:37,595 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testExportExpiredSnapshot 2025-01-08T10:47:37,595 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testExportExpiredSnapshot Metrics about Tables on a single HBase RegionServer 2025-01-08T10:47:37,596 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportExpiredSnapshot 2025-01-08T10:47:37,596 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportExpiredSnapshot Metrics about Tables on a single HBase RegionServer 2025-01-08T10:47:37,597 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T10:47:37,597 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T10:47:43,098 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T10:47:43,873 INFO [Time-limited test {}] snapshot.TestExportSnapshot(468): HDFS export destination path: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333263873 2025-01-08T10:47:43,873 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=hdfs://localhost:37495, tgtDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333263873, rawTgtDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333263873, srcFsUri=hdfs://localhost:37495, srcDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:47:43,902 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:37495, inputRoot=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:47:43,902 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-832679797_22, ugi=jenkins (auth:SIMPLE)]], outputRoot=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333263873, skipTmp=false, initialOutputSnapshotDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333263873/.hbase-snapshot/.tmp/snapshot-testExportExpiredSnapshot 2025-01-08T10:47:43,904 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T10:47:43,905 ERROR [Time-limited test {}] util.AbstractHBaseTool(153): Error running command-line tool org.apache.hadoop.hbase.snapshot.SnapshotTTLExpiredException: TTL for snapshot 'snapshot-testExportExpiredSnapshot' has already expired. at org.apache.hadoop.hbase.snapshot.ExportSnapshot.verifySnapshot(ExportSnapshot.java:948) ~[classes/:?] at org.apache.hadoop.hbase.snapshot.ExportSnapshot.doWork(ExportSnapshot.java:1093) ~[classes/:?] at org.apache.hadoop.hbase.util.AbstractHBaseTool.run(AbstractHBaseTool.java:151) ~[classes/:?] at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:82) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hbase.snapshot.TestExportSnapshot.runExportSnapshot(TestExportSnapshot.java:523) ~[test-classes/:?] at org.apache.hadoop.hbase.snapshot.TestExportSnapshot.testExportExpiredSnapshot(TestExportSnapshot.java:315) ~[test-classes/:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) ~[junit-4.13.2.jar:4.13.2] at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) ~[junit-4.13.2.jar:4.13.2] at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2025-01-08T10:47:43,906 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testExportExpiredSnapshot 2025-01-08T10:47:43,907 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.2 disable testtb-testExportExpiredSnapshot 2025-01-08T10:47:43,907 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=155, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testExportExpiredSnapshot 2025-01-08T10:47:43,909 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=155 2025-01-08T10:47:43,909 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportExpiredSnapshot","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333263909"}]},"ts":"1736333263909"} 2025-01-08T10:47:43,910 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportExpiredSnapshot, state=DISABLING in hbase:meta 2025-01-08T10:47:43,929 INFO [PEWorker-3 {}] procedure.DisableTableProcedure(284): Set testtb-testExportExpiredSnapshot to state=DISABLING 2025-01-08T10:47:43,930 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=156, ppid=155, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testExportExpiredSnapshot}] 2025-01-08T10:47:43,932 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=157, ppid=156, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=85afd80183d737be3864ee9b52caa712, UNASSIGN}, {pid=158, ppid=156, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=97a34fb309b891785b6b66cdc5d5f74b, UNASSIGN}] 2025-01-08T10:47:43,932 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=158, ppid=156, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=97a34fb309b891785b6b66cdc5d5f74b, UNASSIGN 2025-01-08T10:47:43,932 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=157, ppid=156, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=85afd80183d737be3864ee9b52caa712, UNASSIGN 2025-01-08T10:47:43,933 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=158 updating hbase:meta row=97a34fb309b891785b6b66cdc5d5f74b, regionState=CLOSING, regionLocation=67f42ee69b4e,45003,1736333108039 2025-01-08T10:47:43,933 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=157 updating hbase:meta row=85afd80183d737be3864ee9b52caa712, regionState=CLOSING, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:43,934 DEBUG [PEWorker-5 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T10:47:43,934 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=159, ppid=158, state=RUNNABLE; CloseRegionProcedure 97a34fb309b891785b6b66cdc5d5f74b, server=67f42ee69b4e,45003,1736333108039}] 2025-01-08T10:47:43,935 DEBUG [PEWorker-2 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T10:47:43,935 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=160, ppid=157, state=RUNNABLE; CloseRegionProcedure 85afd80183d737be3864ee9b52caa712, server=67f42ee69b4e,36283,1736333108166}] 2025-01-08T10:47:44,010 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=155 2025-01-08T10:47:44,086 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:47:44,086 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:44,086 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] handler.UnassignRegionHandler(124): Close 97a34fb309b891785b6b66cdc5d5f74b 2025-01-08T10:47:44,086 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] handler.UnassignRegionHandler(124): Close 85afd80183d737be3864ee9b52caa712 2025-01-08T10:47:44,086 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T10:47:44,086 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T10:47:44,087 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] regionserver.HRegion(1681): Closing 85afd80183d737be3864ee9b52caa712, disabling compactions & flushes 2025-01-08T10:47:44,087 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] regionserver.HRegion(1681): Closing 97a34fb309b891785b6b66cdc5d5f74b, disabling compactions & flushes 2025-01-08T10:47:44,087 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] regionserver.HRegion(1703): Closing region testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712. 2025-01-08T10:47:44,087 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] regionserver.HRegion(1703): Closing region testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b. 2025-01-08T10:47:44,087 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b. 2025-01-08T10:47:44,087 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712. 2025-01-08T10:47:44,087 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b. after waiting 0 ms 2025-01-08T10:47:44,087 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712. after waiting 0 ms 2025-01-08T10:47:44,087 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b. 2025-01-08T10:47:44,087 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712. 2025-01-08T10:47:44,091 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/97a34fb309b891785b6b66cdc5d5f74b/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T10:47:44,091 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/85afd80183d737be3864ee9b52caa712/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T10:47:44,091 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:47:44,091 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:47:44,091 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] regionserver.HRegion(1922): Closed testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b. 2025-01-08T10:47:44,091 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] regionserver.HRegion(1635): Region close journal for 97a34fb309b891785b6b66cdc5d5f74b: 2025-01-08T10:47:44,091 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] regionserver.HRegion(1922): Closed testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712. 2025-01-08T10:47:44,091 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] regionserver.HRegion(1635): Region close journal for 85afd80183d737be3864ee9b52caa712: 2025-01-08T10:47:44,092 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] handler.UnassignRegionHandler(170): Closed 97a34fb309b891785b6b66cdc5d5f74b 2025-01-08T10:47:44,093 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=158 updating hbase:meta row=97a34fb309b891785b6b66cdc5d5f74b, regionState=CLOSED 2025-01-08T10:47:44,093 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] handler.UnassignRegionHandler(170): Closed 85afd80183d737be3864ee9b52caa712 2025-01-08T10:47:44,094 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=157 updating hbase:meta row=85afd80183d737be3864ee9b52caa712, regionState=CLOSED 2025-01-08T10:47:44,095 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=159, resume processing ppid=158 2025-01-08T10:47:44,095 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=159, ppid=158, state=SUCCESS; CloseRegionProcedure 97a34fb309b891785b6b66cdc5d5f74b, server=67f42ee69b4e,45003,1736333108039 in 160 msec 2025-01-08T10:47:44,096 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=160, resume processing ppid=157 2025-01-08T10:47:44,096 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=160, ppid=157, state=SUCCESS; CloseRegionProcedure 85afd80183d737be3864ee9b52caa712, server=67f42ee69b4e,36283,1736333108166 in 160 msec 2025-01-08T10:47:44,097 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=158, ppid=156, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=97a34fb309b891785b6b66cdc5d5f74b, UNASSIGN in 163 msec 2025-01-08T10:47:44,098 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=157, resume processing ppid=156 2025-01-08T10:47:44,098 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=157, ppid=156, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=85afd80183d737be3864ee9b52caa712, UNASSIGN in 164 msec 2025-01-08T10:47:44,099 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=156, resume processing ppid=155 2025-01-08T10:47:44,099 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=156, ppid=155, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testExportExpiredSnapshot in 168 msec 2025-01-08T10:47:44,100 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportExpiredSnapshot","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333264100"}]},"ts":"1736333264100"} 2025-01-08T10:47:44,101 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportExpiredSnapshot, state=DISABLED in hbase:meta 2025-01-08T10:47:44,112 INFO [PEWorker-4 {}] procedure.DisableTableProcedure(296): Set testtb-testExportExpiredSnapshot to state=DISABLED 2025-01-08T10:47:44,114 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=155, state=SUCCESS; DisableTableProcedure table=testtb-testExportExpiredSnapshot in 206 msec 2025-01-08T10:47:44,211 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=155 2025-01-08T10:47:44,211 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testExportExpiredSnapshot, procId: 155 completed 2025-01-08T10:47:44,211 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.2 delete testtb-testExportExpiredSnapshot 2025-01-08T10:47:44,212 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=161, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testExportExpiredSnapshot 2025-01-08T10:47:44,213 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=161, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportExpiredSnapshot 2025-01-08T10:47:44,213 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testExportExpiredSnapshot 2025-01-08T10:47:44,213 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=161, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testExportExpiredSnapshot 2025-01-08T10:47:44,215 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36283 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testExportExpiredSnapshot 2025-01-08T10:47:44,216 DEBUG [HFileArchiver-16 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/85afd80183d737be3864ee9b52caa712 2025-01-08T10:47:44,216 DEBUG [HFileArchiver-17 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/97a34fb309b891785b6b66cdc5d5f74b 2025-01-08T10:47:44,218 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportExpiredSnapshot 2025-01-08T10:47:44,218 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportExpiredSnapshot 2025-01-08T10:47:44,218 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportExpiredSnapshot 2025-01-08T10:47:44,218 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportExpiredSnapshot 2025-01-08T10:47:44,219 DEBUG [HFileArchiver-16 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/85afd80183d737be3864ee9b52caa712/cf, FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/85afd80183d737be3864ee9b52caa712/recovered.edits] 2025-01-08T10:47:44,219 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF 2025-01-08T10:47:44,219 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF 2025-01-08T10:47:44,219 DEBUG [HFileArchiver-17 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/97a34fb309b891785b6b66cdc5d5f74b/cf, FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/97a34fb309b891785b6b66cdc5d5f74b/recovered.edits] 2025-01-08T10:47:44,220 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportExpiredSnapshot 2025-01-08T10:47:44,220 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:44,220 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportExpiredSnapshot 2025-01-08T10:47:44,220 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:44,220 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:44,220 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data null 2025-01-08T10:47:44,220 INFO [zk-permission-watcher-pool-0 {}] access.AuthManager(135): Skipping permission cache refresh because writable data is empty 2025-01-08T10:47:44,220 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:44,220 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data null 2025-01-08T10:47:44,220 INFO [zk-permission-watcher-pool-0 {}] access.AuthManager(135): Skipping permission cache refresh because writable data is empty 2025-01-08T10:47:44,220 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=161 2025-01-08T10:47:44,221 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:44,221 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:44,221 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:44,222 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:44,223 DEBUG [HFileArchiver-17 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/97a34fb309b891785b6b66cdc5d5f74b/cf/6a68e14d37d14e47a8305cff873d55fe to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportExpiredSnapshot/97a34fb309b891785b6b66cdc5d5f74b/cf/6a68e14d37d14e47a8305cff873d55fe 2025-01-08T10:47:44,223 DEBUG [HFileArchiver-16 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/85afd80183d737be3864ee9b52caa712/cf/63eee5a987bc4a5ab241a37776a04aed to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportExpiredSnapshot/85afd80183d737be3864ee9b52caa712/cf/63eee5a987bc4a5ab241a37776a04aed 2025-01-08T10:47:44,226 DEBUG [HFileArchiver-17 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/97a34fb309b891785b6b66cdc5d5f74b/recovered.edits/9.seqid to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportExpiredSnapshot/97a34fb309b891785b6b66cdc5d5f74b/recovered.edits/9.seqid 2025-01-08T10:47:44,226 DEBUG [HFileArchiver-16 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/85afd80183d737be3864ee9b52caa712/recovered.edits/9.seqid to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportExpiredSnapshot/85afd80183d737be3864ee9b52caa712/recovered.edits/9.seqid 2025-01-08T10:47:44,226 DEBUG [HFileArchiver-17 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/97a34fb309b891785b6b66cdc5d5f74b 2025-01-08T10:47:44,227 DEBUG [HFileArchiver-16 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportExpiredSnapshot/85afd80183d737be3864ee9b52caa712 2025-01-08T10:47:44,227 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(313): Archived testtb-testExportExpiredSnapshot regions 2025-01-08T10:47:44,227 DEBUG [PEWorker-5 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f 2025-01-08T10:47:44,228 DEBUG [PEWorker-5 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf] 2025-01-08T10:47:44,231 DEBUG [PEWorker-5 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/c4ca4238a0b923820dcc509a6f75849b20250108dcdf324001a04f5ba9431dc262de604d_97a34fb309b891785b6b66cdc5d5f74b to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/c4ca4238a0b923820dcc509a6f75849b20250108dcdf324001a04f5ba9431dc262de604d_97a34fb309b891785b6b66cdc5d5f74b 2025-01-08T10:47:44,232 DEBUG [PEWorker-5 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/d41d8cd98f00b204e9800998ecf8427e202501085524df7972ac45c8a7e12cf1e21547b4_85afd80183d737be3864ee9b52caa712 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/d41d8cd98f00b204e9800998ecf8427e202501085524df7972ac45c8a7e12cf1e21547b4_85afd80183d737be3864ee9b52caa712 2025-01-08T10:47:44,232 DEBUG [PEWorker-5 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f 2025-01-08T10:47:44,234 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=161, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testExportExpiredSnapshot 2025-01-08T10:47:44,236 WARN [PEWorker-5 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testtb-testExportExpiredSnapshot from hbase:meta 2025-01-08T10:47:44,239 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testExportExpiredSnapshot' descriptor. 2025-01-08T10:47:44,239 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=161, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testExportExpiredSnapshot 2025-01-08T10:47:44,240 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testExportExpiredSnapshot' from region states. 2025-01-08T10:47:44,240 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736333264240"}]},"ts":"9223372036854775807"} 2025-01-08T10:47:44,240 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736333264240"}]},"ts":"9223372036854775807"} 2025-01-08T10:47:44,241 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T10:47:44,241 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => 85afd80183d737be3864ee9b52caa712, NAME => 'testtb-testExportExpiredSnapshot,,1736333250510.85afd80183d737be3864ee9b52caa712.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => 97a34fb309b891785b6b66cdc5d5f74b, NAME => 'testtb-testExportExpiredSnapshot,1,1736333250510.97a34fb309b891785b6b66cdc5d5f74b.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T10:47:44,242 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testExportExpiredSnapshot' as deleted. 2025-01-08T10:47:44,242 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportExpiredSnapshot","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736333264242"}]},"ts":"9223372036854775807"} 2025-01-08T10:47:44,243 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testExportExpiredSnapshot state from META 2025-01-08T10:47:44,245 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(133): Finished pid=161, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportExpiredSnapshot 2025-01-08T10:47:44,246 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=161, state=SUCCESS; DeleteTableProcedure table=testtb-testExportExpiredSnapshot in 34 msec 2025-01-08T10:47:44,321 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=161 2025-01-08T10:47:44,322 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testExportExpiredSnapshot, procId: 161 completed 2025-01-08T10:47:44,329 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.2 delete name: "emptySnaptb0-testExportExpiredSnapshot" 2025-01-08T10:47:44,331 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(380): Deleting snapshot: emptySnaptb0-testExportExpiredSnapshot 2025-01-08T10:47:44,332 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.2 delete name: "snapshot-testExportExpiredSnapshot" 2025-01-08T10:47:44,334 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(380): Deleting snapshot: snapshot-testExportExpiredSnapshot 2025-01-08T10:47:44,335 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.2 delete name: "snaptb0-testExportExpiredSnapshot" 2025-01-08T10:47:44,337 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testExportExpiredSnapshot 2025-01-08T10:47:44,359 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: snapshot.TestMobSecureExportSnapshot#testExportExpiredSnapshot Thread=793 (was 799), OpenFileDescriptor=795 (was 815), MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=723 (was 875), ProcessCount=13 (was 16), AvailableMemoryMB=1834 (was 1497) - AvailableMemoryMB LEAK? - 2025-01-08T10:47:44,359 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=793 is superior to 500 2025-01-08T10:47:44,384 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: snapshot.TestMobSecureExportSnapshot#testEmptyExportFileSystemState Thread=793, OpenFileDescriptor=795, MaxFileDescriptor=1048576, SystemLoadAverage=723, ProcessCount=13, AvailableMemoryMB=1830 2025-01-08T10:47:44,384 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=793 is superior to 500 2025-01-08T10:47:44,386 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.2 create 'testtb-testEmptyExportFileSystemState', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T10:47:44,387 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=162, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testEmptyExportFileSystemState 2025-01-08T10:47:44,388 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=162, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testEmptyExportFileSystemState execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T10:47:44,388 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "testtb-testEmptyExportFileSystemState" procId is: 162 2025-01-08T10:47:44,389 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=162, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testEmptyExportFileSystemState execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T10:47:44,389 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=162 2025-01-08T10:47:44,399 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742222_1398 (size=448) 2025-01-08T10:47:44,399 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742222_1398 (size=448) 2025-01-08T10:47:44,400 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742222_1398 (size=448) 2025-01-08T10:47:44,402 INFO [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 3ce5869408efc39728677e060fb31c7b, NAME => 'testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testtb-testEmptyExportFileSystemState', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:47:44,403 INFO [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => a988bcbb3b4918c75bfe771ea23570c1, NAME => 'testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testtb-testEmptyExportFileSystemState', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:47:44,414 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742223_1399 (size=73) 2025-01-08T10:47:44,414 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742223_1399 (size=73) 2025-01-08T10:47:44,415 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742223_1399 (size=73) 2025-01-08T10:47:44,415 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:47:44,416 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-1 {}] regionserver.HRegion(1681): Closing a988bcbb3b4918c75bfe771ea23570c1, disabling compactions & flushes 2025-01-08T10:47:44,416 INFO [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1. 2025-01-08T10:47:44,416 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1. 2025-01-08T10:47:44,416 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1. after waiting 0 ms 2025-01-08T10:47:44,416 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1. 2025-01-08T10:47:44,416 INFO [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1. 2025-01-08T10:47:44,416 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-1 {}] regionserver.HRegion(1635): Region close journal for a988bcbb3b4918c75bfe771ea23570c1: 2025-01-08T10:47:44,421 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742224_1400 (size=73) 2025-01-08T10:47:44,421 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742224_1400 (size=73) 2025-01-08T10:47:44,421 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742224_1400 (size=73) 2025-01-08T10:47:44,422 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:47:44,422 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-0 {}] regionserver.HRegion(1681): Closing 3ce5869408efc39728677e060fb31c7b, disabling compactions & flushes 2025-01-08T10:47:44,422 INFO [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b. 2025-01-08T10:47:44,422 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b. 2025-01-08T10:47:44,422 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b. after waiting 0 ms 2025-01-08T10:47:44,422 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b. 2025-01-08T10:47:44,422 INFO [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b. 2025-01-08T10:47:44,423 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-0 {}] regionserver.HRegion(1635): Region close journal for 3ce5869408efc39728677e060fb31c7b: 2025-01-08T10:47:44,424 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=162, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testEmptyExportFileSystemState execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T10:47:44,424 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1.","families":{"info":[{"qualifier":"regioninfo","vlen":72,"tag":[],"timestamp":"1736333264424"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736333264424"}]},"ts":"1736333264424"} 2025-01-08T10:47:44,424 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b.","families":{"info":[{"qualifier":"regioninfo","vlen":72,"tag":[],"timestamp":"1736333264424"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736333264424"}]},"ts":"1736333264424"} 2025-01-08T10:47:44,426 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T10:47:44,427 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=162, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testEmptyExportFileSystemState execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T10:47:44,427 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testEmptyExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333264427"}]},"ts":"1736333264427"} 2025-01-08T10:47:44,428 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testEmptyExportFileSystemState, state=ENABLING in hbase:meta 2025-01-08T10:47:44,432 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(202): Hosts are {67f42ee69b4e=0} racks are {/default-rack=0} 2025-01-08T10:47:44,433 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T10:47:44,433 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T10:47:44,433 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T10:47:44,433 INFO [PEWorker-3 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T10:47:44,433 INFO [PEWorker-3 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T10:47:44,433 INFO [PEWorker-3 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T10:47:44,433 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T10:47:44,433 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=163, ppid=162, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=3ce5869408efc39728677e060fb31c7b, ASSIGN}, {pid=164, ppid=162, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=a988bcbb3b4918c75bfe771ea23570c1, ASSIGN}] 2025-01-08T10:47:44,435 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=164, ppid=162, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=a988bcbb3b4918c75bfe771ea23570c1, ASSIGN 2025-01-08T10:47:44,435 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=163, ppid=162, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=3ce5869408efc39728677e060fb31c7b, ASSIGN 2025-01-08T10:47:44,435 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=163, ppid=162, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=3ce5869408efc39728677e060fb31c7b, ASSIGN; state=OFFLINE, location=67f42ee69b4e,36283,1736333108166; forceNewPlan=false, retain=false 2025-01-08T10:47:44,435 INFO [PEWorker-1 {}] assignment.TransitRegionStateProcedure(264): Starting pid=164, ppid=162, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=a988bcbb3b4918c75bfe771ea23570c1, ASSIGN; state=OFFLINE, location=67f42ee69b4e,38323,1736333108259; forceNewPlan=false, retain=false 2025-01-08T10:47:44,490 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=162 2025-01-08T10:47:44,586 INFO [67f42ee69b4e:34551 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T10:47:44,586 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=163 updating hbase:meta row=3ce5869408efc39728677e060fb31c7b, regionState=OPENING, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:44,586 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=164 updating hbase:meta row=a988bcbb3b4918c75bfe771ea23570c1, regionState=OPENING, regionLocation=67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:44,587 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=165, ppid=164, state=RUNNABLE; OpenRegionProcedure a988bcbb3b4918c75bfe771ea23570c1, server=67f42ee69b4e,38323,1736333108259}] 2025-01-08T10:47:44,588 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=166, ppid=163, state=RUNNABLE; OpenRegionProcedure 3ce5869408efc39728677e060fb31c7b, server=67f42ee69b4e,36283,1736333108166}] 2025-01-08T10:47:44,691 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=162 2025-01-08T10:47:44,739 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:44,739 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:44,742 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] handler.AssignRegionHandler(135): Open testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1. 2025-01-08T10:47:44,742 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] handler.AssignRegionHandler(135): Open testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b. 2025-01-08T10:47:44,742 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(7285): Opening region: {ENCODED => a988bcbb3b4918c75bfe771ea23570c1, NAME => 'testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T10:47:44,742 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(7285): Opening region: {ENCODED => 3ce5869408efc39728677e060fb31c7b, NAME => 'testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T10:47:44,742 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1. service=AccessControlService 2025-01-08T10:47:44,742 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b. service=AccessControlService 2025-01-08T10:47:44,742 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:47:44,742 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:47:44,742 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testEmptyExportFileSystemState a988bcbb3b4918c75bfe771ea23570c1 2025-01-08T10:47:44,743 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testEmptyExportFileSystemState 3ce5869408efc39728677e060fb31c7b 2025-01-08T10:47:44,743 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(894): Instantiated testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:47:44,743 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(894): Instantiated testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:47:44,743 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(7327): checking encryption for 3ce5869408efc39728677e060fb31c7b 2025-01-08T10:47:44,743 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(7327): checking encryption for a988bcbb3b4918c75bfe771ea23570c1 2025-01-08T10:47:44,743 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(7330): checking classloading for 3ce5869408efc39728677e060fb31c7b 2025-01-08T10:47:44,743 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(7330): checking classloading for a988bcbb3b4918c75bfe771ea23570c1 2025-01-08T10:47:44,744 INFO [StoreOpener-a988bcbb3b4918c75bfe771ea23570c1-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region a988bcbb3b4918c75bfe771ea23570c1 2025-01-08T10:47:44,744 INFO [StoreOpener-3ce5869408efc39728677e060fb31c7b-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 3ce5869408efc39728677e060fb31c7b 2025-01-08T10:47:44,745 INFO [StoreOpener-3ce5869408efc39728677e060fb31c7b-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 3ce5869408efc39728677e060fb31c7b columnFamilyName cf 2025-01-08T10:47:44,745 INFO [StoreOpener-a988bcbb3b4918c75bfe771ea23570c1-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region a988bcbb3b4918c75bfe771ea23570c1 columnFamilyName cf 2025-01-08T10:47:44,746 DEBUG [StoreOpener-a988bcbb3b4918c75bfe771ea23570c1-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:44,746 DEBUG [StoreOpener-3ce5869408efc39728677e060fb31c7b-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:44,746 INFO [StoreOpener-3ce5869408efc39728677e060fb31c7b-1 {}] regionserver.HStore(327): Store=3ce5869408efc39728677e060fb31c7b/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:47:44,746 INFO [StoreOpener-a988bcbb3b4918c75bfe771ea23570c1-1 {}] regionserver.HStore(327): Store=a988bcbb3b4918c75bfe771ea23570c1/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:47:44,747 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/3ce5869408efc39728677e060fb31c7b 2025-01-08T10:47:44,747 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/a988bcbb3b4918c75bfe771ea23570c1 2025-01-08T10:47:44,748 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/3ce5869408efc39728677e060fb31c7b 2025-01-08T10:47:44,748 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/a988bcbb3b4918c75bfe771ea23570c1 2025-01-08T10:47:44,749 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(1085): writing seq id for 3ce5869408efc39728677e060fb31c7b 2025-01-08T10:47:44,749 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(1085): writing seq id for a988bcbb3b4918c75bfe771ea23570c1 2025-01-08T10:47:44,751 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/3ce5869408efc39728677e060fb31c7b/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:47:44,751 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/a988bcbb3b4918c75bfe771ea23570c1/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:47:44,751 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(1102): Opened 3ce5869408efc39728677e060fb31c7b; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=60058457, jitterRate=-0.10505925118923187}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T10:47:44,751 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(1102): Opened a988bcbb3b4918c75bfe771ea23570c1; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=74255407, jitterRate=0.10649178922176361}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T10:47:44,752 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(1001): Region open journal for a988bcbb3b4918c75bfe771ea23570c1: 2025-01-08T10:47:44,752 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(1001): Region open journal for 3ce5869408efc39728677e060fb31c7b: 2025-01-08T10:47:44,753 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1., pid=165, masterSystemTime=1736333264739 2025-01-08T10:47:44,753 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b., pid=166, masterSystemTime=1736333264739 2025-01-08T10:47:44,754 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1. 2025-01-08T10:47:44,754 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] handler.AssignRegionHandler(164): Opened testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1. 2025-01-08T10:47:44,754 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=164 updating hbase:meta row=a988bcbb3b4918c75bfe771ea23570c1, regionState=OPEN, openSeqNum=2, regionLocation=67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:44,754 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b. 2025-01-08T10:47:44,755 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] handler.AssignRegionHandler(164): Opened testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b. 2025-01-08T10:47:44,755 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=163 updating hbase:meta row=3ce5869408efc39728677e060fb31c7b, regionState=OPEN, openSeqNum=2, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:44,757 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=165, resume processing ppid=164 2025-01-08T10:47:44,757 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=165, ppid=164, state=SUCCESS; OpenRegionProcedure a988bcbb3b4918c75bfe771ea23570c1, server=67f42ee69b4e,38323,1736333108259 in 169 msec 2025-01-08T10:47:44,758 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=166, resume processing ppid=163 2025-01-08T10:47:44,758 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=166, ppid=163, state=SUCCESS; OpenRegionProcedure 3ce5869408efc39728677e060fb31c7b, server=67f42ee69b4e,36283,1736333108166 in 168 msec 2025-01-08T10:47:44,759 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=164, ppid=162, state=SUCCESS; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=a988bcbb3b4918c75bfe771ea23570c1, ASSIGN in 324 msec 2025-01-08T10:47:44,760 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=163, resume processing ppid=162 2025-01-08T10:47:44,760 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=163, ppid=162, state=SUCCESS; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=3ce5869408efc39728677e060fb31c7b, ASSIGN in 325 msec 2025-01-08T10:47:44,760 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=162, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testEmptyExportFileSystemState execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T10:47:44,760 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testEmptyExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333264760"}]},"ts":"1736333264760"} 2025-01-08T10:47:44,761 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testEmptyExportFileSystemState, state=ENABLED in hbase:meta 2025-01-08T10:47:44,764 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=162, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testEmptyExportFileSystemState execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T10:47:44,764 DEBUG [PEWorker-2 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testEmptyExportFileSystemState jenkins: RWXCA 2025-01-08T10:47:44,765 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36283 {}] access.PermissionStorage(611): Read acl: entry[testtb-testEmptyExportFileSystemState], kv [jenkins: RWXCA] 2025-01-08T10:47:44,772 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:44,772 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:44,772 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:44,772 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:44,774 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testEmptyExportFileSystemState with data PBUF\x0AK\x0A\x07jenkins\x12@\x08\x03"<\x0A0\x0A\x07default\x12%testtb-testEmptyExportFileSystemState \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:44,775 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testEmptyExportFileSystemState with data PBUF\x0AK\x0A\x07jenkins\x12@\x08\x03"<\x0A0\x0A\x07default\x12%testtb-testEmptyExportFileSystemState \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:44,775 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testEmptyExportFileSystemState with data PBUF\x0AK\x0A\x07jenkins\x12@\x08\x03"<\x0A0\x0A\x07default\x12%testtb-testEmptyExportFileSystemState \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:44,775 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:44,775 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:44,775 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:44,775 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testEmptyExportFileSystemState with data PBUF\x0AK\x0A\x07jenkins\x12@\x08\x03"<\x0A0\x0A\x07default\x12%testtb-testEmptyExportFileSystemState \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:44,775 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:44,776 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=162, state=SUCCESS; CreateTableProcedure table=testtb-testEmptyExportFileSystemState in 388 msec 2025-01-08T10:47:44,992 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=162 2025-01-08T10:47:44,992 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testEmptyExportFileSystemState, procId: 162 completed 2025-01-08T10:47:44,994 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testEmptyExportFileSystemState 2025-01-08T10:47:44,994 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b. 2025-01-08T10:47:44,994 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T10:47:45,005 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.2 snapshot request for:{ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } 2025-01-08T10:47:45,005 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736333265005 (current time:1736333265005). 2025-01-08T10:47:45,005 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T10:47:45,005 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot emptySnaptb0-testEmptyExportFileSystemState VERSION not specified, setting to 2 2025-01-08T10:47:45,005 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T10:47:45,006 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x422a6dec to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@4c1823c1 2025-01-08T10:47:45,009 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2234aefe, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:47:45,010 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:47:45,011 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:41100, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:47:45,012 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x422a6dec to 127.0.0.1:52367 2025-01-08T10:47:45,012 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:47:45,012 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x0fd315b4 to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@25829a4 2025-01-08T10:47:45,015 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@59c344f, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:47:45,017 DEBUG [hconnection-0x214b219f-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:47:45,018 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:41116, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:47:45,020 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x0fd315b4 to 127.0.0.1:52367 2025-01-08T10:47:45,020 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:47:45,020 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] access.PermissionStorage(611): Read acl: entry[testtb-testEmptyExportFileSystemState], kv [jenkins: RWXCA] 2025-01-08T10:47:45,021 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T10:47:45,021 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=167, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } 2025-01-08T10:47:45,022 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 }, snapshot procedure id = 167 2025-01-08T10:47:45,022 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T10:47:45,022 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=167 2025-01-08T10:47:45,023 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T10:47:45,025 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T10:47:45,033 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742225_1401 (size=185) 2025-01-08T10:47:45,033 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742225_1401 (size=185) 2025-01-08T10:47:45,033 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742225_1401 (size=185) 2025-01-08T10:47:45,034 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T10:47:45,034 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=168, ppid=167, state=RUNNABLE; SnapshotRegionProcedure 3ce5869408efc39728677e060fb31c7b}, {pid=169, ppid=167, state=RUNNABLE; SnapshotRegionProcedure a988bcbb3b4918c75bfe771ea23570c1}] 2025-01-08T10:47:45,035 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=168, ppid=167, state=RUNNABLE; SnapshotRegionProcedure 3ce5869408efc39728677e060fb31c7b 2025-01-08T10:47:45,035 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=169, ppid=167, state=RUNNABLE; SnapshotRegionProcedure a988bcbb3b4918c75bfe771ea23570c1 2025-01-08T10:47:45,123 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=167 2025-01-08T10:47:45,186 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:45,186 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:45,186 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=38323 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=169 2025-01-08T10:47:45,186 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=36283 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=168 2025-01-08T10:47:45,187 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=168}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b. 2025-01-08T10:47:45,187 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=169}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1. 2025-01-08T10:47:45,187 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=168}] regionserver.HRegion(2538): Flush status journal for 3ce5869408efc39728677e060fb31c7b: 2025-01-08T10:47:45,187 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=169}] regionserver.HRegion(2538): Flush status journal for a988bcbb3b4918c75bfe771ea23570c1: 2025-01-08T10:47:45,187 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=169}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1. for emptySnaptb0-testEmptyExportFileSystemState completed. 2025-01-08T10:47:45,187 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=168}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b. for emptySnaptb0-testEmptyExportFileSystemState completed. 2025-01-08T10:47:45,187 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=168}] snapshot.SnapshotManifest(241): Storing 'testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b.' region-info for snapshot=emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T10:47:45,187 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=168}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:47:45,187 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=169}] snapshot.SnapshotManifest(241): Storing 'testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1.' region-info for snapshot=emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T10:47:45,187 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=169}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:47:45,187 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=168}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T10:47:45,187 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=169}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T10:47:45,200 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742226_1402 (size=76) 2025-01-08T10:47:45,201 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742226_1402 (size=76) 2025-01-08T10:47:45,201 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742226_1402 (size=76) 2025-01-08T10:47:45,201 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=168}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b. 2025-01-08T10:47:45,201 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=168}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=168 2025-01-08T10:47:45,202 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=168 2025-01-08T10:47:45,202 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testEmptyExportFileSystemState on region 3ce5869408efc39728677e060fb31c7b 2025-01-08T10:47:45,202 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=168, ppid=167, state=RUNNABLE; SnapshotRegionProcedure 3ce5869408efc39728677e060fb31c7b 2025-01-08T10:47:45,204 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=168, ppid=167, state=SUCCESS; SnapshotRegionProcedure 3ce5869408efc39728677e060fb31c7b in 169 msec 2025-01-08T10:47:45,210 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742227_1403 (size=76) 2025-01-08T10:47:45,210 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742227_1403 (size=76) 2025-01-08T10:47:45,210 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742227_1403 (size=76) 2025-01-08T10:47:45,211 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=169}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1. 2025-01-08T10:47:45,211 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=169}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=169 2025-01-08T10:47:45,211 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=169 2025-01-08T10:47:45,211 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testEmptyExportFileSystemState on region a988bcbb3b4918c75bfe771ea23570c1 2025-01-08T10:47:45,212 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=169, ppid=167, state=RUNNABLE; SnapshotRegionProcedure a988bcbb3b4918c75bfe771ea23570c1 2025-01-08T10:47:45,214 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=169, resume processing ppid=167 2025-01-08T10:47:45,214 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T10:47:45,214 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=169, ppid=167, state=SUCCESS; SnapshotRegionProcedure a988bcbb3b4918c75bfe771ea23570c1 in 178 msec 2025-01-08T10:47:45,214 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T10:47:45,215 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T10:47:45,215 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T10:47:45,216 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:45,216 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(328): No files under family: cf 2025-01-08T10:47:45,222 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742228_1404 (size=68) 2025-01-08T10:47:45,222 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742228_1404 (size=68) 2025-01-08T10:47:45,223 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742228_1404 (size=68) 2025-01-08T10:47:45,224 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T10:47:45,224 DEBUG [PEWorker-5 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T10:47:45,224 DEBUG [PEWorker-5 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T10:47:45,238 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742229_1405 (size=673) 2025-01-08T10:47:45,238 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742229_1405 (size=673) 2025-01-08T10:47:45,239 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742229_1405 (size=673) 2025-01-08T10:47:45,241 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T10:47:45,245 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T10:47:45,245 DEBUG [PEWorker-5 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/emptySnaptb0-testEmptyExportFileSystemState to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T10:47:45,247 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T10:47:45,247 DEBUG [PEWorker-5 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 }, snapshot procedure id = 167 2025-01-08T10:47:45,248 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=167, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } in 226 msec 2025-01-08T10:47:45,324 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=167 2025-01-08T10:47:45,324 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testEmptyExportFileSystemState, procId: 167 completed 2025-01-08T10:47:45,326 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36283 {}] regionserver.HRegion(8254): writing data to region testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T10:47:45,327 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38323 {}] regionserver.HRegion(8254): writing data to region testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T10:47:45,329 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testEmptyExportFileSystemState 2025-01-08T10:47:45,329 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b. 2025-01-08T10:47:45,329 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T10:47:45,338 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.2 snapshot request for:{ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } 2025-01-08T10:47:45,338 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736333265338 (current time:1736333265338). 2025-01-08T10:47:45,338 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T10:47:45,338 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testEmptyExportFileSystemState VERSION not specified, setting to 2 2025-01-08T10:47:45,338 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T10:47:45,339 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x11ae8863 to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@69111410 2025-01-08T10:47:45,342 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@486798c3, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:47:45,343 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:47:45,344 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:41122, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:47:45,345 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x11ae8863 to 127.0.0.1:52367 2025-01-08T10:47:45,345 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:47:45,346 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x7f6fa9ba to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@1ebf93d3 2025-01-08T10:47:45,348 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@6a6aa147, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:47:45,350 DEBUG [hconnection-0x4e110c70-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:47:45,351 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:41134, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:47:45,353 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x7f6fa9ba to 127.0.0.1:52367 2025-01-08T10:47:45,353 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:47:45,353 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] access.PermissionStorage(611): Read acl: entry[testtb-testEmptyExportFileSystemState], kv [jenkins: RWXCA] 2025-01-08T10:47:45,353 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T10:47:45,354 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=170, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } 2025-01-08T10:47:45,354 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 }, snapshot procedure id = 170 2025-01-08T10:47:45,355 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=170 2025-01-08T10:47:45,355 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T10:47:45,356 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T10:47:45,358 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T10:47:45,363 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742230_1406 (size=180) 2025-01-08T10:47:45,363 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742230_1406 (size=180) 2025-01-08T10:47:45,363 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742230_1406 (size=180) 2025-01-08T10:47:45,364 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T10:47:45,364 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=171, ppid=170, state=RUNNABLE; SnapshotRegionProcedure 3ce5869408efc39728677e060fb31c7b}, {pid=172, ppid=170, state=RUNNABLE; SnapshotRegionProcedure a988bcbb3b4918c75bfe771ea23570c1}] 2025-01-08T10:47:45,365 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=172, ppid=170, state=RUNNABLE; SnapshotRegionProcedure a988bcbb3b4918c75bfe771ea23570c1 2025-01-08T10:47:45,365 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=171, ppid=170, state=RUNNABLE; SnapshotRegionProcedure 3ce5869408efc39728677e060fb31c7b 2025-01-08T10:47:45,456 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=170 2025-01-08T10:47:45,516 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:45,516 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:45,516 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=36283 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=171 2025-01-08T10:47:45,516 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=38323 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=172 2025-01-08T10:47:45,517 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b. 2025-01-08T10:47:45,517 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1. 2025-01-08T10:47:45,517 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.HRegion(2837): Flushing 3ce5869408efc39728677e060fb31c7b 1/1 column families, dataSize=266 B heapSize=832 B 2025-01-08T10:47:45,517 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.HRegion(2837): Flushing a988bcbb3b4918c75bfe771ea23570c1 1/1 column families, dataSize=3.00 KB heapSize=6.72 KB 2025-01-08T10:47:45,535 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108159d45d93a4f41fb982a720e0a321d22_3ce5869408efc39728677e060fb31c7b is 71, key is 02f1fc7db70ca22383a63064db7080fa/cf:q/1736333265326/Put/seqid=0 2025-01-08T10:47:45,535 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108142692cd9b0a428c8e6ee3503243fd6b_a988bcbb3b4918c75bfe771ea23570c1 is 71, key is 10df9682a0643a0b1070e17cd5e0f226/cf:q/1736333265327/Put/seqid=0 2025-01-08T10:47:45,544 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742231_1407 (size=5171) 2025-01-08T10:47:45,544 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742232_1408 (size=8101) 2025-01-08T10:47:45,544 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742232_1408 (size=8101) 2025-01-08T10:47:45,544 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742231_1407 (size=5171) 2025-01-08T10:47:45,544 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742231_1407 (size=5171) 2025-01-08T10:47:45,545 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742232_1408 (size=8101) 2025-01-08T10:47:45,545 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:45,545 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:45,549 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108159d45d93a4f41fb982a720e0a321d22_3ce5869408efc39728677e060fb31c7b to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/d41d8cd98f00b204e9800998ecf8427e20250108159d45d93a4f41fb982a720e0a321d22_3ce5869408efc39728677e060fb31c7b 2025-01-08T10:47:45,549 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108142692cd9b0a428c8e6ee3503243fd6b_a988bcbb3b4918c75bfe771ea23570c1 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/c4ca4238a0b923820dcc509a6f75849b20250108142692cd9b0a428c8e6ee3503243fd6b_a988bcbb3b4918c75bfe771ea23570c1 2025-01-08T10:47:45,550 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/a988bcbb3b4918c75bfe771ea23570c1/.tmp/cf/376e1a34f69e43d4befa3f2777d1cb6a, store: [table=testtb-testEmptyExportFileSystemState family=cf region=a988bcbb3b4918c75bfe771ea23570c1] 2025-01-08T10:47:45,550 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/3ce5869408efc39728677e060fb31c7b/.tmp/cf/687fa0772ce54eb4b930b0c2a3c84d33, store: [table=testtb-testEmptyExportFileSystemState family=cf region=3ce5869408efc39728677e060fb31c7b] 2025-01-08T10:47:45,551 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/a988bcbb3b4918c75bfe771ea23570c1/.tmp/cf/376e1a34f69e43d4befa3f2777d1cb6a is 214, key is 1d1da3d3fec309de700109a471a766c71/cf:q/1736333265327/Put/seqid=0 2025-01-08T10:47:45,551 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/3ce5869408efc39728677e060fb31c7b/.tmp/cf/687fa0772ce54eb4b930b0c2a3c84d33 is 214, key is 06e1fbb37040e392fdbedfebd8cd3dddc/cf:q/1736333265326/Put/seqid=0 2025-01-08T10:47:45,556 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742234_1410 (size=15027) 2025-01-08T10:47:45,556 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742234_1410 (size=15027) 2025-01-08T10:47:45,556 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742233_1409 (size=6146) 2025-01-08T10:47:45,556 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742234_1410 (size=15027) 2025-01-08T10:47:45,556 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742233_1409 (size=6146) 2025-01-08T10:47:45,557 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742233_1409 (size=6146) 2025-01-08T10:47:45,557 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=3.0 K, hasBloomFilter=true, into tmp file hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/a988bcbb3b4918c75bfe771ea23570c1/.tmp/cf/376e1a34f69e43d4befa3f2777d1cb6a 2025-01-08T10:47:45,557 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=266, hasBloomFilter=true, into tmp file hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/3ce5869408efc39728677e060fb31c7b/.tmp/cf/687fa0772ce54eb4b930b0c2a3c84d33 2025-01-08T10:47:45,562 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/a988bcbb3b4918c75bfe771ea23570c1/.tmp/cf/376e1a34f69e43d4befa3f2777d1cb6a as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/a988bcbb3b4918c75bfe771ea23570c1/cf/376e1a34f69e43d4befa3f2777d1cb6a 2025-01-08T10:47:45,562 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/3ce5869408efc39728677e060fb31c7b/.tmp/cf/687fa0772ce54eb4b930b0c2a3c84d33 as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/3ce5869408efc39728677e060fb31c7b/cf/687fa0772ce54eb4b930b0c2a3c84d33 2025-01-08T10:47:45,566 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/a988bcbb3b4918c75bfe771ea23570c1/cf/376e1a34f69e43d4befa3f2777d1cb6a, entries=46, sequenceid=6, filesize=14.7 K 2025-01-08T10:47:45,566 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/3ce5869408efc39728677e060fb31c7b/cf/687fa0772ce54eb4b930b0c2a3c84d33, entries=4, sequenceid=6, filesize=6.0 K 2025-01-08T10:47:45,567 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.HRegion(3040): Finished flush of dataSize ~3.00 KB/3070, heapSize ~6.70 KB/6864, currentSize=0 B/0 for a988bcbb3b4918c75bfe771ea23570c1 in 50ms, sequenceid=6, compaction requested=false 2025-01-08T10:47:45,567 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.HRegion(3040): Finished flush of dataSize ~266 B/266, heapSize ~816 B/816, currentSize=0 B/0 for 3ce5869408efc39728677e060fb31c7b in 50ms, sequenceid=6, compaction requested=false 2025-01-08T10:47:45,567 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testEmptyExportFileSystemState' 2025-01-08T10:47:45,567 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testEmptyExportFileSystemState' 2025-01-08T10:47:45,567 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.HRegion(2538): Flush status journal for a988bcbb3b4918c75bfe771ea23570c1: 2025-01-08T10:47:45,567 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.HRegion(2538): Flush status journal for 3ce5869408efc39728677e060fb31c7b: 2025-01-08T10:47:45,568 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1. for snaptb0-testEmptyExportFileSystemState completed. 2025-01-08T10:47:45,568 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b. for snaptb0-testEmptyExportFileSystemState completed. 2025-01-08T10:47:45,568 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] snapshot.SnapshotManifest(241): Storing 'testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1.' region-info for snapshot=snaptb0-testEmptyExportFileSystemState 2025-01-08T10:47:45,568 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] snapshot.SnapshotManifest(241): Storing 'testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b.' region-info for snapshot=snaptb0-testEmptyExportFileSystemState 2025-01-08T10:47:45,568 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:47:45,568 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:47:45,568 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/a988bcbb3b4918c75bfe771ea23570c1/cf/376e1a34f69e43d4befa3f2777d1cb6a] hfiles 2025-01-08T10:47:45,568 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/3ce5869408efc39728677e060fb31c7b/cf/687fa0772ce54eb4b930b0c2a3c84d33] hfiles 2025-01-08T10:47:45,568 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/a988bcbb3b4918c75bfe771ea23570c1/cf/376e1a34f69e43d4befa3f2777d1cb6a for snapshot=snaptb0-testEmptyExportFileSystemState 2025-01-08T10:47:45,568 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/3ce5869408efc39728677e060fb31c7b/cf/687fa0772ce54eb4b930b0c2a3c84d33 for snapshot=snaptb0-testEmptyExportFileSystemState 2025-01-08T10:47:45,580 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742235_1411 (size=115) 2025-01-08T10:47:45,580 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742235_1411 (size=115) 2025-01-08T10:47:45,580 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742235_1411 (size=115) 2025-01-08T10:47:45,581 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1. 2025-01-08T10:47:45,581 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=172 2025-01-08T10:47:45,581 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=172 2025-01-08T10:47:45,581 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testEmptyExportFileSystemState on region a988bcbb3b4918c75bfe771ea23570c1 2025-01-08T10:47:45,581 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=172, ppid=170, state=RUNNABLE; SnapshotRegionProcedure a988bcbb3b4918c75bfe771ea23570c1 2025-01-08T10:47:45,583 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=172, ppid=170, state=SUCCESS; SnapshotRegionProcedure a988bcbb3b4918c75bfe771ea23570c1 in 218 msec 2025-01-08T10:47:45,591 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742236_1412 (size=115) 2025-01-08T10:47:45,591 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742236_1412 (size=115) 2025-01-08T10:47:45,592 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742236_1412 (size=115) 2025-01-08T10:47:45,592 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b. 2025-01-08T10:47:45,592 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=171 2025-01-08T10:47:45,592 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=171 2025-01-08T10:47:45,592 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testEmptyExportFileSystemState on region 3ce5869408efc39728677e060fb31c7b 2025-01-08T10:47:45,592 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=171, ppid=170, state=RUNNABLE; SnapshotRegionProcedure 3ce5869408efc39728677e060fb31c7b 2025-01-08T10:47:45,595 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=171, resume processing ppid=170 2025-01-08T10:47:45,595 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=171, ppid=170, state=SUCCESS; SnapshotRegionProcedure 3ce5869408efc39728677e060fb31c7b in 229 msec 2025-01-08T10:47:45,595 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T10:47:45,595 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T10:47:45,596 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T10:47:45,596 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T10:47:45,596 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:45,598 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/c4ca4238a0b923820dcc509a6f75849b20250108142692cd9b0a428c8e6ee3503243fd6b_a988bcbb3b4918c75bfe771ea23570c1, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/d41d8cd98f00b204e9800998ecf8427e20250108159d45d93a4f41fb982a720e0a321d22_3ce5869408efc39728677e060fb31c7b] hfiles 2025-01-08T10:47:45,598 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/c4ca4238a0b923820dcc509a6f75849b20250108142692cd9b0a428c8e6ee3503243fd6b_a988bcbb3b4918c75bfe771ea23570c1 2025-01-08T10:47:45,598 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/d41d8cd98f00b204e9800998ecf8427e20250108159d45d93a4f41fb982a720e0a321d22_3ce5869408efc39728677e060fb31c7b 2025-01-08T10:47:45,603 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742237_1413 (size=299) 2025-01-08T10:47:45,603 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742237_1413 (size=299) 2025-01-08T10:47:45,603 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742237_1413 (size=299) 2025-01-08T10:47:45,605 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T10:47:45,605 DEBUG [PEWorker-5 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testEmptyExportFileSystemState 2025-01-08T10:47:45,606 DEBUG [PEWorker-5 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/snaptb0-testEmptyExportFileSystemState 2025-01-08T10:47:45,616 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742238_1414 (size=983) 2025-01-08T10:47:45,616 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742238_1414 (size=983) 2025-01-08T10:47:45,617 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742238_1414 (size=983) 2025-01-08T10:47:45,622 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T10:47:45,628 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T10:47:45,628 DEBUG [PEWorker-5 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/snaptb0-testEmptyExportFileSystemState to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testEmptyExportFileSystemState 2025-01-08T10:47:45,629 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T10:47:45,630 DEBUG [PEWorker-5 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 }, snapshot procedure id = 170 2025-01-08T10:47:45,631 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=170, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } in 277 msec 2025-01-08T10:47:45,657 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=170 2025-01-08T10:47:45,657 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testEmptyExportFileSystemState, procId: 170 completed 2025-01-08T10:47:45,657 INFO [Time-limited test {}] snapshot.TestExportSnapshot(468): HDFS export destination path: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333265657 2025-01-08T10:47:45,657 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=hdfs://localhost:37495, tgtDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333265657, rawTgtDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333265657, srcFsUri=hdfs://localhost:37495, srcDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:47:45,687 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:37495, inputRoot=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:47:45,687 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-832679797_22, ugi=jenkins (auth:SIMPLE)]], outputRoot=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333265657, skipTmp=false, initialOutputSnapshotDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333265657/.hbase-snapshot/.tmp/emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T10:47:45,689 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T10:47:45,693 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333265657/.hbase-snapshot/.tmp/emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T10:47:45,711 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742239_1415 (size=673) 2025-01-08T10:47:45,711 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742239_1415 (size=673) 2025-01-08T10:47:45,711 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742239_1415 (size=673) 2025-01-08T10:47:45,716 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742240_1416 (size=185) 2025-01-08T10:47:45,716 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742240_1416 (size=185) 2025-01-08T10:47:45,716 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742240_1416 (size=185) 2025-01-08T10:47:45,889 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-8489290506569642287.jar 2025-01-08T10:47:45,890 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:45,891 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:45,891 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:46,892 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-16853851660139714865.jar 2025-01-08T10:47:46,892 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:46,892 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:46,961 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-4567759671253148101.jar 2025-01-08T10:47:46,961 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:46,961 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:46,962 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:46,962 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:46,962 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:46,962 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:46,962 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T10:47:46,963 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T10:47:46,963 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T10:47:46,963 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T10:47:46,963 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T10:47:46,963 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T10:47:46,964 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T10:47:46,964 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T10:47:46,964 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T10:47:46,964 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T10:47:46,964 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T10:47:46,964 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T10:47:46,965 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:47:46,965 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:47:46,965 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:47:46,965 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:47:46,966 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:47:46,966 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:47:46,966 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:47:47,036 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742241_1417 (size=127628) 2025-01-08T10:47:47,036 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742241_1417 (size=127628) 2025-01-08T10:47:47,037 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742241_1417 (size=127628) 2025-01-08T10:47:47,055 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742242_1418 (size=2172137) 2025-01-08T10:47:47,055 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742242_1418 (size=2172137) 2025-01-08T10:47:47,055 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742242_1418 (size=2172137) 2025-01-08T10:47:47,063 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742243_1419 (size=213228) 2025-01-08T10:47:47,063 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742243_1419 (size=213228) 2025-01-08T10:47:47,063 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742243_1419 (size=213228) 2025-01-08T10:47:47,075 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742244_1420 (size=1877034) 2025-01-08T10:47:47,075 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742244_1420 (size=1877034) 2025-01-08T10:47:47,075 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742244_1420 (size=1877034) 2025-01-08T10:47:47,083 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742245_1421 (size=533455) 2025-01-08T10:47:47,083 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742245_1421 (size=533455) 2025-01-08T10:47:47,084 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742245_1421 (size=533455) 2025-01-08T10:47:47,106 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742246_1422 (size=6350714) 2025-01-08T10:47:47,107 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742246_1422 (size=6350714) 2025-01-08T10:47:47,107 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742246_1422 (size=6350714) 2025-01-08T10:47:47,132 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742247_1423 (size=7280644) 2025-01-08T10:47:47,132 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742247_1423 (size=7280644) 2025-01-08T10:47:47,132 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742247_1423 (size=7280644) 2025-01-08T10:47:47,149 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742248_1424 (size=4188619) 2025-01-08T10:47:47,149 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742248_1424 (size=4188619) 2025-01-08T10:47:47,149 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742248_1424 (size=4188619) 2025-01-08T10:47:47,156 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742249_1425 (size=20406) 2025-01-08T10:47:47,157 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742249_1425 (size=20406) 2025-01-08T10:47:47,157 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742249_1425 (size=20406) 2025-01-08T10:47:47,168 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742250_1426 (size=75495) 2025-01-08T10:47:47,168 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742250_1426 (size=75495) 2025-01-08T10:47:47,168 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742250_1426 (size=75495) 2025-01-08T10:47:47,175 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742251_1427 (size=45609) 2025-01-08T10:47:47,175 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742251_1427 (size=45609) 2025-01-08T10:47:47,175 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742251_1427 (size=45609) 2025-01-08T10:47:47,185 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742252_1428 (size=110084) 2025-01-08T10:47:47,186 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742252_1428 (size=110084) 2025-01-08T10:47:47,186 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742252_1428 (size=110084) 2025-01-08T10:47:47,195 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742253_1429 (size=451756) 2025-01-08T10:47:47,195 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742253_1429 (size=451756) 2025-01-08T10:47:47,195 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742253_1429 (size=451756) 2025-01-08T10:47:47,205 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742254_1430 (size=1323991) 2025-01-08T10:47:47,205 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742254_1430 (size=1323991) 2025-01-08T10:47:47,206 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742254_1430 (size=1323991) 2025-01-08T10:47:47,216 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742255_1431 (size=23076) 2025-01-08T10:47:47,216 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742255_1431 (size=23076) 2025-01-08T10:47:47,216 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742255_1431 (size=23076) 2025-01-08T10:47:47,224 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742256_1432 (size=126803) 2025-01-08T10:47:47,224 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742256_1432 (size=126803) 2025-01-08T10:47:47,224 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742256_1432 (size=126803) 2025-01-08T10:47:47,235 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742257_1433 (size=322274) 2025-01-08T10:47:47,235 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742257_1433 (size=322274) 2025-01-08T10:47:47,236 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742257_1433 (size=322274) 2025-01-08T10:47:47,249 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742258_1434 (size=1832290) 2025-01-08T10:47:47,249 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742258_1434 (size=1832290) 2025-01-08T10:47:47,250 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742258_1434 (size=1832290) 2025-01-08T10:47:47,258 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742259_1435 (size=30081) 2025-01-08T10:47:47,258 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742259_1435 (size=30081) 2025-01-08T10:47:47,258 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742259_1435 (size=30081) 2025-01-08T10:47:47,265 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742260_1436 (size=53616) 2025-01-08T10:47:47,265 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742260_1436 (size=53616) 2025-01-08T10:47:47,266 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742260_1436 (size=53616) 2025-01-08T10:47:47,277 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742261_1437 (size=29229) 2025-01-08T10:47:47,278 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742261_1437 (size=29229) 2025-01-08T10:47:47,278 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742261_1437 (size=29229) 2025-01-08T10:47:47,285 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742262_1438 (size=169089) 2025-01-08T10:47:47,286 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742262_1438 (size=169089) 2025-01-08T10:47:47,286 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742262_1438 (size=169089) 2025-01-08T10:47:47,296 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742263_1439 (size=912094) 2025-01-08T10:47:47,296 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742263_1439 (size=912094) 2025-01-08T10:47:47,297 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742263_1439 (size=912094) 2025-01-08T10:47:47,329 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742264_1440 (size=5175431) 2025-01-08T10:47:47,329 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742264_1440 (size=5175431) 2025-01-08T10:47:47,330 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742264_1440 (size=5175431) 2025-01-08T10:47:47,337 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742265_1441 (size=136454) 2025-01-08T10:47:47,338 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742265_1441 (size=136454) 2025-01-08T10:47:47,338 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742265_1441 (size=136454) 2025-01-08T10:47:47,354 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742266_1442 (size=3317408) 2025-01-08T10:47:47,354 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742266_1442 (size=3317408) 2025-01-08T10:47:47,355 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742266_1442 (size=3317408) 2025-01-08T10:47:47,363 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742267_1443 (size=503880) 2025-01-08T10:47:47,364 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742267_1443 (size=503880) 2025-01-08T10:47:47,364 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742267_1443 (size=503880) 2025-01-08T10:47:47,389 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742268_1444 (size=4695811) 2025-01-08T10:47:47,390 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742268_1444 (size=4695811) 2025-01-08T10:47:47,390 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742268_1444 (size=4695811) 2025-01-08T10:47:47,392 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T10:47:47,394 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'emptySnaptb0-testEmptyExportFileSystemState' hfile list 2025-01-08T10:47:47,401 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742269_1445 (size=7) 2025-01-08T10:47:47,401 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742269_1445 (size=7) 2025-01-08T10:47:47,402 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742269_1445 (size=7) 2025-01-08T10:47:47,407 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742270_1446 (size=10) 2025-01-08T10:47:47,407 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742270_1446 (size=10) 2025-01-08T10:47:47,407 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742270_1446 (size=10) 2025-01-08T10:47:47,419 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742271_1447 (size=304832) 2025-01-08T10:47:47,419 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742271_1447 (size=304832) 2025-01-08T10:47:47,420 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742271_1447 (size=304832) 2025-01-08T10:47:47,436 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T10:47:47,436 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T10:47:47,549 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0007_000001 (auth:SIMPLE) from 127.0.0.1:40312 2025-01-08T10:47:47,596 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testEmptyExportFileSystemState 2025-01-08T10:47:47,596 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testEmptyExportFileSystemState Metrics about Tables on a single HBase RegionServer 2025-01-08T10:47:47,596 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportExpiredSnapshot 2025-01-08T10:47:49,593 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T10:47:52,440 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0007_000001 (auth:SIMPLE) from 127.0.0.1:41676 2025-01-08T10:47:52,696 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742272_1448 (size=350482) 2025-01-08T10:47:52,696 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742272_1448 (size=350482) 2025-01-08T10:47:52,697 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742272_1448 (size=350482) 2025-01-08T10:47:53,531 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742273_1449 (size=8568) 2025-01-08T10:47:53,531 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742273_1449 (size=8568) 2025-01-08T10:47:53,531 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742273_1449 (size=8568) 2025-01-08T10:47:53,544 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742274_1450 (size=460) 2025-01-08T10:47:53,544 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742274_1450 (size=460) 2025-01-08T10:47:53,544 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742274_1450 (size=460) 2025-01-08T10:47:53,572 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742275_1451 (size=8568) 2025-01-08T10:47:53,572 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742275_1451 (size=8568) 2025-01-08T10:47:53,573 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742275_1451 (size=8568) 2025-01-08T10:47:53,594 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742276_1452 (size=350482) 2025-01-08T10:47:53,595 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742276_1452 (size=350482) 2025-01-08T10:47:53,595 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742276_1452 (size=350482) 2025-01-08T10:47:54,750 INFO [Time-limited test {}] snapshot.ExportSnapshot(1207): Finalize the Snapshot Export 2025-01-08T10:47:54,751 INFO [Time-limited test {}] snapshot.ExportSnapshot(1218): Verify the exported snapshot's expiration status and integrity. 2025-01-08T10:47:54,755 INFO [Time-limited test {}] snapshot.ExportSnapshot(1224): Export Completed: emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T10:47:54,755 INFO [Time-limited test {}] snapshot.TestExportSnapshot(362): Exported snapshot 2025-01-08T10:47:54,756 INFO [Time-limited test {}] snapshot.TestExportSnapshot(373): Verified filesystem state 2025-01-08T10:47:54,756 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-832679797_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState at hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T10:47:54,756 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState/.snapshotinfo 2025-01-08T10:47:54,756 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState/data.manifest 2025-01-08T10:47:54,756 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-832679797_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333265657/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState at hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333265657/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T10:47:54,757 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333265657/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState/.snapshotinfo 2025-01-08T10:47:54,757 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333265657/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState/data.manifest 2025-01-08T10:47:54,761 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testEmptyExportFileSystemState 2025-01-08T10:47:54,761 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.2 disable testtb-testEmptyExportFileSystemState 2025-01-08T10:47:54,762 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=173, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testEmptyExportFileSystemState 2025-01-08T10:47:54,763 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=173 2025-01-08T10:47:54,764 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testEmptyExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333274764"}]},"ts":"1736333274764"} 2025-01-08T10:47:54,765 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testEmptyExportFileSystemState, state=DISABLING in hbase:meta 2025-01-08T10:47:54,767 INFO [PEWorker-1 {}] procedure.DisableTableProcedure(284): Set testtb-testEmptyExportFileSystemState to state=DISABLING 2025-01-08T10:47:54,768 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=174, ppid=173, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testEmptyExportFileSystemState}] 2025-01-08T10:47:54,769 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=175, ppid=174, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=3ce5869408efc39728677e060fb31c7b, UNASSIGN}, {pid=176, ppid=174, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=a988bcbb3b4918c75bfe771ea23570c1, UNASSIGN}] 2025-01-08T10:47:54,769 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=176, ppid=174, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=a988bcbb3b4918c75bfe771ea23570c1, UNASSIGN 2025-01-08T10:47:54,770 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=175, ppid=174, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=3ce5869408efc39728677e060fb31c7b, UNASSIGN 2025-01-08T10:47:54,770 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=176 updating hbase:meta row=a988bcbb3b4918c75bfe771ea23570c1, regionState=CLOSING, regionLocation=67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:54,770 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=175 updating hbase:meta row=3ce5869408efc39728677e060fb31c7b, regionState=CLOSING, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:54,771 DEBUG [PEWorker-3 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T10:47:54,771 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=177, ppid=175, state=RUNNABLE; CloseRegionProcedure 3ce5869408efc39728677e060fb31c7b, server=67f42ee69b4e,36283,1736333108166}] 2025-01-08T10:47:54,772 DEBUG [PEWorker-2 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T10:47:54,772 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=178, ppid=176, state=RUNNABLE; CloseRegionProcedure a988bcbb3b4918c75bfe771ea23570c1, server=67f42ee69b4e,38323,1736333108259}] 2025-01-08T10:47:54,864 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=173 2025-01-08T10:47:54,923 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:54,923 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] handler.UnassignRegionHandler(124): Close 3ce5869408efc39728677e060fb31c7b 2025-01-08T10:47:54,923 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T10:47:54,924 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:54,924 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] regionserver.HRegion(1681): Closing 3ce5869408efc39728677e060fb31c7b, disabling compactions & flushes 2025-01-08T10:47:54,924 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] regionserver.HRegion(1703): Closing region testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b. 2025-01-08T10:47:54,924 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b. 2025-01-08T10:47:54,924 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] regionserver.HRegion(1791): Acquired close lock on testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b. after waiting 0 ms 2025-01-08T10:47:54,924 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] regionserver.HRegion(1801): Updates disabled for region testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b. 2025-01-08T10:47:54,924 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] handler.UnassignRegionHandler(124): Close a988bcbb3b4918c75bfe771ea23570c1 2025-01-08T10:47:54,924 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T10:47:54,924 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] regionserver.HRegion(1681): Closing a988bcbb3b4918c75bfe771ea23570c1, disabling compactions & flushes 2025-01-08T10:47:54,924 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] regionserver.HRegion(1703): Closing region testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1. 2025-01-08T10:47:54,924 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1. 2025-01-08T10:47:54,924 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] regionserver.HRegion(1791): Acquired close lock on testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1. after waiting 0 ms 2025-01-08T10:47:54,924 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] regionserver.HRegion(1801): Updates disabled for region testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1. 2025-01-08T10:47:54,928 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/3ce5869408efc39728677e060fb31c7b/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T10:47:54,929 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/a988bcbb3b4918c75bfe771ea23570c1/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T10:47:54,929 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:47:54,929 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] regionserver.HRegion(1922): Closed testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b. 2025-01-08T10:47:54,929 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] regionserver.HRegion(1635): Region close journal for 3ce5869408efc39728677e060fb31c7b: 2025-01-08T10:47:54,929 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:47:54,929 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] regionserver.HRegion(1922): Closed testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1. 2025-01-08T10:47:54,929 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] regionserver.HRegion(1635): Region close journal for a988bcbb3b4918c75bfe771ea23570c1: 2025-01-08T10:47:54,930 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] handler.UnassignRegionHandler(170): Closed 3ce5869408efc39728677e060fb31c7b 2025-01-08T10:47:54,931 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=175 updating hbase:meta row=3ce5869408efc39728677e060fb31c7b, regionState=CLOSED 2025-01-08T10:47:54,931 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] handler.UnassignRegionHandler(170): Closed a988bcbb3b4918c75bfe771ea23570c1 2025-01-08T10:47:54,931 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=176 updating hbase:meta row=a988bcbb3b4918c75bfe771ea23570c1, regionState=CLOSED 2025-01-08T10:47:54,934 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=177, resume processing ppid=175 2025-01-08T10:47:54,935 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=178, resume processing ppid=176 2025-01-08T10:47:54,935 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=177, ppid=175, state=SUCCESS; CloseRegionProcedure 3ce5869408efc39728677e060fb31c7b, server=67f42ee69b4e,36283,1736333108166 in 161 msec 2025-01-08T10:47:54,935 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=175, ppid=174, state=SUCCESS; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=3ce5869408efc39728677e060fb31c7b, UNASSIGN in 165 msec 2025-01-08T10:47:54,935 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=178, ppid=176, state=SUCCESS; CloseRegionProcedure a988bcbb3b4918c75bfe771ea23570c1, server=67f42ee69b4e,38323,1736333108259 in 161 msec 2025-01-08T10:47:54,936 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=176, resume processing ppid=174 2025-01-08T10:47:54,936 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=176, ppid=174, state=SUCCESS; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=a988bcbb3b4918c75bfe771ea23570c1, UNASSIGN in 166 msec 2025-01-08T10:47:54,938 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=174, resume processing ppid=173 2025-01-08T10:47:54,938 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=174, ppid=173, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testEmptyExportFileSystemState in 169 msec 2025-01-08T10:47:54,939 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testEmptyExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333274939"}]},"ts":"1736333274939"} 2025-01-08T10:47:54,940 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testEmptyExportFileSystemState, state=DISABLED in hbase:meta 2025-01-08T10:47:54,941 INFO [PEWorker-4 {}] procedure.DisableTableProcedure(296): Set testtb-testEmptyExportFileSystemState to state=DISABLED 2025-01-08T10:47:54,943 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=173, state=SUCCESS; DisableTableProcedure table=testtb-testEmptyExportFileSystemState in 181 msec 2025-01-08T10:47:55,065 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=173 2025-01-08T10:47:55,065 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testEmptyExportFileSystemState, procId: 173 completed 2025-01-08T10:47:55,066 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.2 delete testtb-testEmptyExportFileSystemState 2025-01-08T10:47:55,066 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=179, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testEmptyExportFileSystemState 2025-01-08T10:47:55,067 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=179, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testEmptyExportFileSystemState 2025-01-08T10:47:55,067 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testEmptyExportFileSystemState 2025-01-08T10:47:55,068 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=179, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testEmptyExportFileSystemState 2025-01-08T10:47:55,069 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36283 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testEmptyExportFileSystemState 2025-01-08T10:47:55,070 DEBUG [HFileArchiver-12 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/a988bcbb3b4918c75bfe771ea23570c1 2025-01-08T10:47:55,070 DEBUG [HFileArchiver-18 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/3ce5869408efc39728677e060fb31c7b 2025-01-08T10:47:55,072 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testEmptyExportFileSystemState 2025-01-08T10:47:55,072 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testEmptyExportFileSystemState 2025-01-08T10:47:55,072 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testEmptyExportFileSystemState 2025-01-08T10:47:55,073 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testEmptyExportFileSystemState 2025-01-08T10:47:55,073 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testEmptyExportFileSystemState with data PBUF 2025-01-08T10:47:55,073 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testEmptyExportFileSystemState with data PBUF 2025-01-08T10:47:55,073 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testEmptyExportFileSystemState with data PBUF 2025-01-08T10:47:55,073 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testEmptyExportFileSystemState with data PBUF 2025-01-08T10:47:55,075 DEBUG [HFileArchiver-18 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/3ce5869408efc39728677e060fb31c7b/cf, FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/3ce5869408efc39728677e060fb31c7b/recovered.edits] 2025-01-08T10:47:55,075 DEBUG [HFileArchiver-12 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/a988bcbb3b4918c75bfe771ea23570c1/cf, FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/a988bcbb3b4918c75bfe771ea23570c1/recovered.edits] 2025-01-08T10:47:55,076 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testEmptyExportFileSystemState 2025-01-08T10:47:55,076 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testEmptyExportFileSystemState 2025-01-08T10:47:55,076 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:55,076 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:55,076 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testEmptyExportFileSystemState 2025-01-08T10:47:55,076 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testEmptyExportFileSystemState 2025-01-08T10:47:55,077 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:55,077 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:55,078 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=179 2025-01-08T10:47:55,078 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:55,078 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:55,078 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:55,078 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:55,080 DEBUG [HFileArchiver-12 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/a988bcbb3b4918c75bfe771ea23570c1/cf/376e1a34f69e43d4befa3f2777d1cb6a to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testEmptyExportFileSystemState/a988bcbb3b4918c75bfe771ea23570c1/cf/376e1a34f69e43d4befa3f2777d1cb6a 2025-01-08T10:47:55,081 DEBUG [HFileArchiver-18 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/3ce5869408efc39728677e060fb31c7b/cf/687fa0772ce54eb4b930b0c2a3c84d33 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testEmptyExportFileSystemState/3ce5869408efc39728677e060fb31c7b/cf/687fa0772ce54eb4b930b0c2a3c84d33 2025-01-08T10:47:55,084 DEBUG [HFileArchiver-12 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/a988bcbb3b4918c75bfe771ea23570c1/recovered.edits/9.seqid to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testEmptyExportFileSystemState/a988bcbb3b4918c75bfe771ea23570c1/recovered.edits/9.seqid 2025-01-08T10:47:55,085 DEBUG [HFileArchiver-12 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/a988bcbb3b4918c75bfe771ea23570c1 2025-01-08T10:47:55,085 DEBUG [HFileArchiver-18 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/3ce5869408efc39728677e060fb31c7b/recovered.edits/9.seqid to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testEmptyExportFileSystemState/3ce5869408efc39728677e060fb31c7b/recovered.edits/9.seqid 2025-01-08T10:47:55,085 DEBUG [HFileArchiver-18 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testEmptyExportFileSystemState/3ce5869408efc39728677e060fb31c7b 2025-01-08T10:47:55,085 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(313): Archived testtb-testEmptyExportFileSystemState regions 2025-01-08T10:47:55,086 DEBUG [PEWorker-2 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0 2025-01-08T10:47:55,087 DEBUG [PEWorker-2 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf] 2025-01-08T10:47:55,090 DEBUG [PEWorker-2 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/c4ca4238a0b923820dcc509a6f75849b20250108142692cd9b0a428c8e6ee3503243fd6b_a988bcbb3b4918c75bfe771ea23570c1 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/c4ca4238a0b923820dcc509a6f75849b20250108142692cd9b0a428c8e6ee3503243fd6b_a988bcbb3b4918c75bfe771ea23570c1 2025-01-08T10:47:55,091 DEBUG [PEWorker-2 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/d41d8cd98f00b204e9800998ecf8427e20250108159d45d93a4f41fb982a720e0a321d22_3ce5869408efc39728677e060fb31c7b to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/d41d8cd98f00b204e9800998ecf8427e20250108159d45d93a4f41fb982a720e0a321d22_3ce5869408efc39728677e060fb31c7b 2025-01-08T10:47:55,092 DEBUG [PEWorker-2 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0 2025-01-08T10:47:55,097 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=179, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testEmptyExportFileSystemState 2025-01-08T10:47:55,100 WARN [PEWorker-2 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testtb-testEmptyExportFileSystemState from hbase:meta 2025-01-08T10:47:55,102 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testEmptyExportFileSystemState' descriptor. 2025-01-08T10:47:55,102 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=179, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testEmptyExportFileSystemState 2025-01-08T10:47:55,103 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testEmptyExportFileSystemState' from region states. 2025-01-08T10:47:55,103 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736333275103"}]},"ts":"9223372036854775807"} 2025-01-08T10:47:55,103 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736333275103"}]},"ts":"9223372036854775807"} 2025-01-08T10:47:55,105 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T10:47:55,105 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => 3ce5869408efc39728677e060fb31c7b, NAME => 'testtb-testEmptyExportFileSystemState,,1736333264385.3ce5869408efc39728677e060fb31c7b.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => a988bcbb3b4918c75bfe771ea23570c1, NAME => 'testtb-testEmptyExportFileSystemState,1,1736333264385.a988bcbb3b4918c75bfe771ea23570c1.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T10:47:55,105 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testEmptyExportFileSystemState' as deleted. 2025-01-08T10:47:55,105 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testEmptyExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736333275105"}]},"ts":"9223372036854775807"} 2025-01-08T10:47:55,106 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testEmptyExportFileSystemState state from META 2025-01-08T10:47:55,108 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(133): Finished pid=179, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testEmptyExportFileSystemState 2025-01-08T10:47:55,109 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=179, state=SUCCESS; DeleteTableProcedure table=testtb-testEmptyExportFileSystemState in 42 msec 2025-01-08T10:47:55,179 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=179 2025-01-08T10:47:55,179 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testEmptyExportFileSystemState, procId: 179 completed 2025-01-08T10:47:55,184 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.2 delete name: "emptySnaptb0-testEmptyExportFileSystemState" 2025-01-08T10:47:55,186 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(380): Deleting snapshot: emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T10:47:55,187 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.2 delete name: "snaptb0-testEmptyExportFileSystemState" 2025-01-08T10:47:55,188 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testEmptyExportFileSystemState 2025-01-08T10:47:55,208 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: snapshot.TestMobSecureExportSnapshot#testEmptyExportFileSystemState Thread=806 (was 793) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-832679797_22 at /127.0.0.1:55044 [Waiting for operation #5] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (175901922) connection to localhost/127.0.0.1:45269 from appattempt_1736333116683_0007_000001 java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_1660468606_1 at /127.0.0.1:33100 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-41 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-42 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (175901922) connection to localhost/127.0.0.1:38333 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: Thread-5989 java.base@17.0.11/java.io.FileInputStream.readBytes(Native Method) java.base@17.0.11/java.io.FileInputStream.read(FileInputStream.java:276) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:282) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) java.base@17.0.11/sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:281) java.base@17.0.11/sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:324) java.base@17.0.11/sun.nio.cs.StreamDecoder.read(StreamDecoder.java:189) java.base@17.0.11/java.io.InputStreamReader.read(InputStreamReader.java:177) java.base@17.0.11/java.io.BufferedReader.fill(BufferedReader.java:162) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:329) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:396) app//org.apache.hadoop.util.Shell$1.run(Shell.java:1025) Potentially hanging thread: HFileArchiver-18 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #12 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-832679797_22 at /127.0.0.1:46170 [Waiting for operation #4] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_1660468606_1 at /127.0.0.1:40950 [Waiting for operation #3] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-832679797_22 at /127.0.0.1:38726 [Waiting for operation #3] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-40 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: process reaper (pid 23912) java.base@17.0.11/java.lang.ProcessHandleImpl.waitForProcessExit0(Native Method) java.base@17.0.11/java.lang.ProcessHandleImpl$1.run(ProcessHandleImpl.java:150) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:38333 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-43 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) - Thread LEAK? -, OpenFileDescriptor=827 (was 795) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=643 (was 723), ProcessCount=18 (was 13) - ProcessCount LEAK? -, AvailableMemoryMB=1415 (was 1830) 2025-01-08T10:47:55,208 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=806 is superior to 500 2025-01-08T10:47:55,228 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: snapshot.TestMobSecureExportSnapshot#testExportWithChecksum Thread=806, OpenFileDescriptor=827, MaxFileDescriptor=1048576, SystemLoadAverage=643, ProcessCount=18, AvailableMemoryMB=1413 2025-01-08T10:47:55,228 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=806 is superior to 500 2025-01-08T10:47:55,229 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.2 create 'testtb-testExportWithChecksum', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T10:47:55,230 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=180, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testExportWithChecksum 2025-01-08T10:47:55,231 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=180, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportWithChecksum execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T10:47:55,231 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "testtb-testExportWithChecksum" procId is: 180 2025-01-08T10:47:55,232 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=180, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testExportWithChecksum execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T10:47:55,232 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=180 2025-01-08T10:47:55,239 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742277_1453 (size=440) 2025-01-08T10:47:55,239 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742277_1453 (size=440) 2025-01-08T10:47:55,240 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742277_1453 (size=440) 2025-01-08T10:47:55,241 INFO [RegionOpenAndInit-testtb-testExportWithChecksum-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 6dd967531e3eb9fd666bb33f76786e5c, NAME => 'testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testtb-testExportWithChecksum', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:47:55,242 INFO [RegionOpenAndInit-testtb-testExportWithChecksum-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => 367c2df29adffa5404639744fd4a5501, NAME => 'testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testtb-testExportWithChecksum', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:47:55,252 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742278_1454 (size=65) 2025-01-08T10:47:55,252 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742278_1454 (size=65) 2025-01-08T10:47:55,252 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742278_1454 (size=65) 2025-01-08T10:47:55,253 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:47:55,253 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-1 {}] regionserver.HRegion(1681): Closing 367c2df29adffa5404639744fd4a5501, disabling compactions & flushes 2025-01-08T10:47:55,253 INFO [RegionOpenAndInit-testtb-testExportWithChecksum-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501. 2025-01-08T10:47:55,254 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501. 2025-01-08T10:47:55,254 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501. after waiting 0 ms 2025-01-08T10:47:55,254 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501. 2025-01-08T10:47:55,254 INFO [RegionOpenAndInit-testtb-testExportWithChecksum-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501. 2025-01-08T10:47:55,254 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-1 {}] regionserver.HRegion(1635): Region close journal for 367c2df29adffa5404639744fd4a5501: 2025-01-08T10:47:55,260 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742279_1455 (size=65) 2025-01-08T10:47:55,260 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742279_1455 (size=65) 2025-01-08T10:47:55,260 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742279_1455 (size=65) 2025-01-08T10:47:55,261 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:47:55,261 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-0 {}] regionserver.HRegion(1681): Closing 6dd967531e3eb9fd666bb33f76786e5c, disabling compactions & flushes 2025-01-08T10:47:55,261 INFO [RegionOpenAndInit-testtb-testExportWithChecksum-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c. 2025-01-08T10:47:55,261 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c. 2025-01-08T10:47:55,261 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c. after waiting 0 ms 2025-01-08T10:47:55,261 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c. 2025-01-08T10:47:55,261 INFO [RegionOpenAndInit-testtb-testExportWithChecksum-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c. 2025-01-08T10:47:55,261 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-0 {}] regionserver.HRegion(1635): Region close journal for 6dd967531e3eb9fd666bb33f76786e5c: 2025-01-08T10:47:55,262 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=180, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testExportWithChecksum execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T10:47:55,262 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501.","families":{"info":[{"qualifier":"regioninfo","vlen":64,"tag":[],"timestamp":"1736333275262"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736333275262"}]},"ts":"1736333275262"} 2025-01-08T10:47:55,262 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c.","families":{"info":[{"qualifier":"regioninfo","vlen":64,"tag":[],"timestamp":"1736333275262"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736333275262"}]},"ts":"1736333275262"} 2025-01-08T10:47:55,265 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T10:47:55,265 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=180, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testExportWithChecksum execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T10:47:55,266 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithChecksum","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333275265"}]},"ts":"1736333275265"} 2025-01-08T10:47:55,267 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithChecksum, state=ENABLING in hbase:meta 2025-01-08T10:47:55,271 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(202): Hosts are {67f42ee69b4e=0} racks are {/default-rack=0} 2025-01-08T10:47:55,272 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T10:47:55,272 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T10:47:55,272 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T10:47:55,272 INFO [PEWorker-3 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T10:47:55,272 INFO [PEWorker-3 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T10:47:55,272 INFO [PEWorker-3 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T10:47:55,272 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T10:47:55,273 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=181, ppid=180, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=6dd967531e3eb9fd666bb33f76786e5c, ASSIGN}, {pid=182, ppid=180, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=367c2df29adffa5404639744fd4a5501, ASSIGN}] 2025-01-08T10:47:55,274 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=182, ppid=180, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=367c2df29adffa5404639744fd4a5501, ASSIGN 2025-01-08T10:47:55,274 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=181, ppid=180, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=6dd967531e3eb9fd666bb33f76786e5c, ASSIGN 2025-01-08T10:47:55,275 INFO [PEWorker-5 {}] assignment.TransitRegionStateProcedure(264): Starting pid=182, ppid=180, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=367c2df29adffa5404639744fd4a5501, ASSIGN; state=OFFLINE, location=67f42ee69b4e,38323,1736333108259; forceNewPlan=false, retain=false 2025-01-08T10:47:55,275 INFO [PEWorker-1 {}] assignment.TransitRegionStateProcedure(264): Starting pid=181, ppid=180, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=6dd967531e3eb9fd666bb33f76786e5c, ASSIGN; state=OFFLINE, location=67f42ee69b4e,36283,1736333108166; forceNewPlan=false, retain=false 2025-01-08T10:47:55,334 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=180 2025-01-08T10:47:55,425 INFO [67f42ee69b4e:34551 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T10:47:55,425 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=181 updating hbase:meta row=6dd967531e3eb9fd666bb33f76786e5c, regionState=OPENING, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:55,425 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=182 updating hbase:meta row=367c2df29adffa5404639744fd4a5501, regionState=OPENING, regionLocation=67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:55,427 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=183, ppid=182, state=RUNNABLE; OpenRegionProcedure 367c2df29adffa5404639744fd4a5501, server=67f42ee69b4e,38323,1736333108259}] 2025-01-08T10:47:55,428 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=184, ppid=181, state=RUNNABLE; OpenRegionProcedure 6dd967531e3eb9fd666bb33f76786e5c, server=67f42ee69b4e,36283,1736333108166}] 2025-01-08T10:47:55,535 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=180 2025-01-08T10:47:55,579 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:55,579 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:55,582 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] handler.AssignRegionHandler(135): Open testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501. 2025-01-08T10:47:55,582 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(7285): Opening region: {ENCODED => 367c2df29adffa5404639744fd4a5501, NAME => 'testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T10:47:55,582 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501. service=AccessControlService 2025-01-08T10:47:55,583 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] handler.AssignRegionHandler(135): Open testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c. 2025-01-08T10:47:55,583 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(7285): Opening region: {ENCODED => 6dd967531e3eb9fd666bb33f76786e5c, NAME => 'testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T10:47:55,583 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:47:55,583 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c. service=AccessControlService 2025-01-08T10:47:55,583 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportWithChecksum 367c2df29adffa5404639744fd4a5501 2025-01-08T10:47:55,583 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:47:55,583 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(894): Instantiated testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:47:55,583 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(7327): checking encryption for 367c2df29adffa5404639744fd4a5501 2025-01-08T10:47:55,583 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportWithChecksum 6dd967531e3eb9fd666bb33f76786e5c 2025-01-08T10:47:55,583 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(7330): checking classloading for 367c2df29adffa5404639744fd4a5501 2025-01-08T10:47:55,583 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(894): Instantiated testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:47:55,583 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(7327): checking encryption for 6dd967531e3eb9fd666bb33f76786e5c 2025-01-08T10:47:55,583 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(7330): checking classloading for 6dd967531e3eb9fd666bb33f76786e5c 2025-01-08T10:47:55,585 INFO [StoreOpener-6dd967531e3eb9fd666bb33f76786e5c-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 6dd967531e3eb9fd666bb33f76786e5c 2025-01-08T10:47:55,586 INFO [StoreOpener-367c2df29adffa5404639744fd4a5501-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 367c2df29adffa5404639744fd4a5501 2025-01-08T10:47:55,586 INFO [StoreOpener-6dd967531e3eb9fd666bb33f76786e5c-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 6dd967531e3eb9fd666bb33f76786e5c columnFamilyName cf 2025-01-08T10:47:55,587 DEBUG [StoreOpener-6dd967531e3eb9fd666bb33f76786e5c-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:55,587 INFO [StoreOpener-367c2df29adffa5404639744fd4a5501-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 367c2df29adffa5404639744fd4a5501 columnFamilyName cf 2025-01-08T10:47:55,588 INFO [StoreOpener-6dd967531e3eb9fd666bb33f76786e5c-1 {}] regionserver.HStore(327): Store=6dd967531e3eb9fd666bb33f76786e5c/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:47:55,588 DEBUG [StoreOpener-367c2df29adffa5404639744fd4a5501-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:55,589 INFO [StoreOpener-367c2df29adffa5404639744fd4a5501-1 {}] regionserver.HStore(327): Store=367c2df29adffa5404639744fd4a5501/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:47:55,589 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/6dd967531e3eb9fd666bb33f76786e5c 2025-01-08T10:47:55,590 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/6dd967531e3eb9fd666bb33f76786e5c 2025-01-08T10:47:55,590 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/367c2df29adffa5404639744fd4a5501 2025-01-08T10:47:55,591 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/367c2df29adffa5404639744fd4a5501 2025-01-08T10:47:55,592 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(1085): writing seq id for 6dd967531e3eb9fd666bb33f76786e5c 2025-01-08T10:47:55,593 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(1085): writing seq id for 367c2df29adffa5404639744fd4a5501 2025-01-08T10:47:55,594 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/6dd967531e3eb9fd666bb33f76786e5c/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:47:55,595 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/367c2df29adffa5404639744fd4a5501/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:47:55,595 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(1102): Opened 6dd967531e3eb9fd666bb33f76786e5c; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=74409839, jitterRate=0.10879300534725189}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T10:47:55,595 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(1001): Region open journal for 6dd967531e3eb9fd666bb33f76786e5c: 2025-01-08T10:47:55,596 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(1102): Opened 367c2df29adffa5404639744fd4a5501; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=67453151, jitterRate=0.005130276083946228}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T10:47:55,596 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(1001): Region open journal for 367c2df29adffa5404639744fd4a5501: 2025-01-08T10:47:55,597 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501., pid=183, masterSystemTime=1736333275579 2025-01-08T10:47:55,597 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c., pid=184, masterSystemTime=1736333275579 2025-01-08T10:47:55,603 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c. 2025-01-08T10:47:55,603 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] handler.AssignRegionHandler(164): Opened testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c. 2025-01-08T10:47:55,603 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=181 updating hbase:meta row=6dd967531e3eb9fd666bb33f76786e5c, regionState=OPEN, openSeqNum=2, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:55,603 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501. 2025-01-08T10:47:55,604 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] handler.AssignRegionHandler(164): Opened testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501. 2025-01-08T10:47:55,604 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=182 updating hbase:meta row=367c2df29adffa5404639744fd4a5501, regionState=OPEN, openSeqNum=2, regionLocation=67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:55,607 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=184, resume processing ppid=181 2025-01-08T10:47:55,607 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=184, ppid=181, state=SUCCESS; OpenRegionProcedure 6dd967531e3eb9fd666bb33f76786e5c, server=67f42ee69b4e,36283,1736333108166 in 177 msec 2025-01-08T10:47:55,608 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=183, resume processing ppid=182 2025-01-08T10:47:55,609 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=183, ppid=182, state=SUCCESS; OpenRegionProcedure 367c2df29adffa5404639744fd4a5501, server=67f42ee69b4e,38323,1736333108259 in 179 msec 2025-01-08T10:47:55,610 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=181, ppid=180, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=6dd967531e3eb9fd666bb33f76786e5c, ASSIGN in 334 msec 2025-01-08T10:47:55,611 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=182, resume processing ppid=180 2025-01-08T10:47:55,611 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=182, ppid=180, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=367c2df29adffa5404639744fd4a5501, ASSIGN in 336 msec 2025-01-08T10:47:55,611 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=180, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testExportWithChecksum execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T10:47:55,611 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithChecksum","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333275611"}]},"ts":"1736333275611"} 2025-01-08T10:47:55,612 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithChecksum, state=ENABLED in hbase:meta 2025-01-08T10:47:55,615 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=180, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportWithChecksum execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T10:47:55,615 DEBUG [PEWorker-1 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testExportWithChecksum jenkins: RWXCA 2025-01-08T10:47:55,617 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36283 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportWithChecksum], kv [jenkins: RWXCA] 2025-01-08T10:47:55,619 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:55,619 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:55,619 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:55,619 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:47:55,621 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:55,621 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:55,621 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithChecksum with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithChecksum \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:55,621 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithChecksum with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithChecksum \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:55,623 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:55,623 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithChecksum with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithChecksum \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:55,624 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:55,624 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithChecksum with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithChecksum \x00 \x01 \x02 \x03 \x04 2025-01-08T10:47:55,625 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=180, state=SUCCESS; CreateTableProcedure table=testtb-testExportWithChecksum in 391 msec 2025-01-08T10:47:55,836 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=180 2025-01-08T10:47:55,836 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testExportWithChecksum, procId: 180 completed 2025-01-08T10:47:55,838 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportWithChecksum 2025-01-08T10:47:55,838 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c. 2025-01-08T10:47:55,839 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T10:47:55,854 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.2 snapshot request for:{ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } 2025-01-08T10:47:55,854 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736333275854 (current time:1736333275854). 2025-01-08T10:47:55,855 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T10:47:55,855 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot emptySnaptb0-testExportWithChecksum VERSION not specified, setting to 2 2025-01-08T10:47:55,855 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T10:47:55,856 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x71be25f1 to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@fee741 2025-01-08T10:47:55,864 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@6ebb34d4, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:47:55,865 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:47:55,866 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:58402, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:47:55,867 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x71be25f1 to 127.0.0.1:52367 2025-01-08T10:47:55,867 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:47:55,868 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x107078ce to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@67c3416e 2025-01-08T10:47:55,876 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5069d4d6, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:47:55,878 DEBUG [hconnection-0x4a263df3-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:47:55,879 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:58414, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:47:55,881 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x107078ce to 127.0.0.1:52367 2025-01-08T10:47:55,881 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:47:55,881 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportWithChecksum], kv [jenkins: RWXCA] 2025-01-08T10:47:55,882 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T10:47:55,883 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=185, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } 2025-01-08T10:47:55,883 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 }, snapshot procedure id = 185 2025-01-08T10:47:55,884 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=185 2025-01-08T10:47:55,884 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T10:47:55,885 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T10:47:55,886 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T10:47:55,903 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742280_1456 (size=161) 2025-01-08T10:47:55,904 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742280_1456 (size=161) 2025-01-08T10:47:55,904 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742280_1456 (size=161) 2025-01-08T10:47:55,905 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T10:47:55,906 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=186, ppid=185, state=RUNNABLE; SnapshotRegionProcedure 6dd967531e3eb9fd666bb33f76786e5c}, {pid=187, ppid=185, state=RUNNABLE; SnapshotRegionProcedure 367c2df29adffa5404639744fd4a5501}] 2025-01-08T10:47:55,906 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=186, ppid=185, state=RUNNABLE; SnapshotRegionProcedure 6dd967531e3eb9fd666bb33f76786e5c 2025-01-08T10:47:55,906 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=187, ppid=185, state=RUNNABLE; SnapshotRegionProcedure 367c2df29adffa5404639744fd4a5501 2025-01-08T10:47:55,985 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=185 2025-01-08T10:47:56,057 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:56,057 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:56,058 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=36283 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=186 2025-01-08T10:47:56,058 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=38323 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=187 2025-01-08T10:47:56,058 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=186}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c. 2025-01-08T10:47:56,058 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=187}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501. 2025-01-08T10:47:56,059 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=186}] regionserver.HRegion(2538): Flush status journal for 6dd967531e3eb9fd666bb33f76786e5c: 2025-01-08T10:47:56,059 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=187}] regionserver.HRegion(2538): Flush status journal for 367c2df29adffa5404639744fd4a5501: 2025-01-08T10:47:56,059 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=186}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c. for emptySnaptb0-testExportWithChecksum completed. 2025-01-08T10:47:56,059 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=187}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501. for emptySnaptb0-testExportWithChecksum completed. 2025-01-08T10:47:56,059 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=186}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c.' region-info for snapshot=emptySnaptb0-testExportWithChecksum 2025-01-08T10:47:56,059 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=187}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501.' region-info for snapshot=emptySnaptb0-testExportWithChecksum 2025-01-08T10:47:56,059 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=186}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:47:56,059 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=187}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:47:56,059 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=186}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T10:47:56,059 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=187}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T10:47:56,075 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742282_1458 (size=68) 2025-01-08T10:47:56,075 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742282_1458 (size=68) 2025-01-08T10:47:56,076 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742282_1458 (size=68) 2025-01-08T10:47:56,076 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=187}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501. 2025-01-08T10:47:56,076 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=187}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=187 2025-01-08T10:47:56,077 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=187 2025-01-08T10:47:56,077 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportWithChecksum on region 367c2df29adffa5404639744fd4a5501 2025-01-08T10:47:56,077 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=187, ppid=185, state=RUNNABLE; SnapshotRegionProcedure 367c2df29adffa5404639744fd4a5501 2025-01-08T10:47:56,079 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=187, ppid=185, state=SUCCESS; SnapshotRegionProcedure 367c2df29adffa5404639744fd4a5501 in 172 msec 2025-01-08T10:47:56,081 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742281_1457 (size=68) 2025-01-08T10:47:56,082 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742281_1457 (size=68) 2025-01-08T10:47:56,082 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742281_1457 (size=68) 2025-01-08T10:47:56,083 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=186}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c. 2025-01-08T10:47:56,083 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=186}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=186 2025-01-08T10:47:56,083 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=186 2025-01-08T10:47:56,083 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportWithChecksum on region 6dd967531e3eb9fd666bb33f76786e5c 2025-01-08T10:47:56,083 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=186, ppid=185, state=RUNNABLE; SnapshotRegionProcedure 6dd967531e3eb9fd666bb33f76786e5c 2025-01-08T10:47:56,086 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=186, resume processing ppid=185 2025-01-08T10:47:56,086 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=186, ppid=185, state=SUCCESS; SnapshotRegionProcedure 6dd967531e3eb9fd666bb33f76786e5c in 178 msec 2025-01-08T10:47:56,086 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T10:47:56,086 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T10:47:56,088 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T10:47:56,088 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T10:47:56,088 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:56,088 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(328): No files under family: cf 2025-01-08T10:47:56,100 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742283_1459 (size=60) 2025-01-08T10:47:56,100 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742283_1459 (size=60) 2025-01-08T10:47:56,101 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742283_1459 (size=60) 2025-01-08T10:47:56,102 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T10:47:56,102 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for emptySnaptb0-testExportWithChecksum 2025-01-08T10:47:56,102 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/emptySnaptb0-testExportWithChecksum 2025-01-08T10:47:56,127 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742284_1460 (size=641) 2025-01-08T10:47:56,127 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742284_1460 (size=641) 2025-01-08T10:47:56,127 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742284_1460 (size=641) 2025-01-08T10:47:56,132 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T10:47:56,138 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T10:47:56,138 DEBUG [PEWorker-1 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/emptySnaptb0-testExportWithChecksum to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/emptySnaptb0-testExportWithChecksum 2025-01-08T10:47:56,140 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T10:47:56,140 DEBUG [PEWorker-1 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 }, snapshot procedure id = 185 2025-01-08T10:47:56,141 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=185, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } in 258 msec 2025-01-08T10:47:56,186 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=185 2025-01-08T10:47:56,186 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportWithChecksum, procId: 185 completed 2025-01-08T10:47:56,189 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36283 {}] regionserver.HRegion(8254): writing data to region testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T10:47:56,193 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38323 {}] regionserver.HRegion(8254): writing data to region testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T10:47:56,196 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportWithChecksum 2025-01-08T10:47:56,196 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c. 2025-01-08T10:47:56,196 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T10:47:56,209 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.2 snapshot request for:{ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } 2025-01-08T10:47:56,209 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736333276209 (current time:1736333276209). 2025-01-08T10:47:56,209 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T10:47:56,209 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testExportWithChecksum VERSION not specified, setting to 2 2025-01-08T10:47:56,209 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T10:47:56,210 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x3a67c73a to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@41d5bc20 2025-01-08T10:47:56,216 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4407669, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:47:56,217 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:47:56,219 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:58418, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:47:56,220 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x3a67c73a to 127.0.0.1:52367 2025-01-08T10:47:56,220 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:47:56,221 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x17fbc33e to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@7ba3495d 2025-01-08T10:47:56,231 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7cd78b7e, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:47:56,233 DEBUG [hconnection-0x63aa56b3-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:47:56,234 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:58428, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:47:56,240 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x17fbc33e to 127.0.0.1:52367 2025-01-08T10:47:56,240 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:47:56,240 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportWithChecksum], kv [jenkins: RWXCA] 2025-01-08T10:47:56,241 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T10:47:56,242 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=188, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } 2025-01-08T10:47:56,242 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 }, snapshot procedure id = 188 2025-01-08T10:47:56,243 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=188 2025-01-08T10:47:56,243 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T10:47:56,244 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T10:47:56,246 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T10:47:56,276 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742285_1461 (size=156) 2025-01-08T10:47:56,276 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742285_1461 (size=156) 2025-01-08T10:47:56,277 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742285_1461 (size=156) 2025-01-08T10:47:56,278 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T10:47:56,278 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=189, ppid=188, state=RUNNABLE; SnapshotRegionProcedure 6dd967531e3eb9fd666bb33f76786e5c}, {pid=190, ppid=188, state=RUNNABLE; SnapshotRegionProcedure 367c2df29adffa5404639744fd4a5501}] 2025-01-08T10:47:56,279 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=189, ppid=188, state=RUNNABLE; SnapshotRegionProcedure 6dd967531e3eb9fd666bb33f76786e5c 2025-01-08T10:47:56,279 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=190, ppid=188, state=RUNNABLE; SnapshotRegionProcedure 367c2df29adffa5404639744fd4a5501 2025-01-08T10:47:56,344 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=188 2025-01-08T10:47:56,430 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:47:56,430 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,38323,1736333108259 2025-01-08T10:47:56,431 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=36283 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=189 2025-01-08T10:47:56,431 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=38323 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=190 2025-01-08T10:47:56,431 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c. 2025-01-08T10:47:56,431 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501. 2025-01-08T10:47:56,431 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.HRegion(2837): Flushing 6dd967531e3eb9fd666bb33f76786e5c 1/1 column families, dataSize=266 B heapSize=832 B 2025-01-08T10:47:56,431 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.HRegion(2837): Flushing 367c2df29adffa5404639744fd4a5501 1/1 column families, dataSize=3.00 KB heapSize=6.72 KB 2025-01-08T10:47:56,457 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108eabde8fd0b394cdaa85e862d9ac36216_6dd967531e3eb9fd666bb33f76786e5c is 71, key is 04ce950e6f5d78155a82986582128046/cf:q/1736333276189/Put/seqid=0 2025-01-08T10:47:56,457 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b202501088931dc000d18485880dd36fdb85181e2_367c2df29adffa5404639744fd4a5501 is 71, key is 147e9c39664c243eccbaabeec5de88ef/cf:q/1736333276192/Put/seqid=0 2025-01-08T10:47:56,472 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742286_1462 (size=5172) 2025-01-08T10:47:56,473 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742286_1462 (size=5172) 2025-01-08T10:47:56,473 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742286_1462 (size=5172) 2025-01-08T10:47:56,473 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:56,478 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108eabde8fd0b394cdaa85e862d9ac36216_6dd967531e3eb9fd666bb33f76786e5c to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/d41d8cd98f00b204e9800998ecf8427e20250108eabde8fd0b394cdaa85e862d9ac36216_6dd967531e3eb9fd666bb33f76786e5c 2025-01-08T10:47:56,479 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/6dd967531e3eb9fd666bb33f76786e5c/.tmp/cf/746ed8a4c2a642458956c576a74fbbbe, store: [table=testtb-testExportWithChecksum family=cf region=6dd967531e3eb9fd666bb33f76786e5c] 2025-01-08T10:47:56,480 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/6dd967531e3eb9fd666bb33f76786e5c/.tmp/cf/746ed8a4c2a642458956c576a74fbbbe is 206, key is 03c4b72f8ec05efbfe3efd8e45cf7ea2f/cf:q/1736333276189/Put/seqid=0 2025-01-08T10:47:56,484 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742287_1463 (size=8101) 2025-01-08T10:47:56,485 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742287_1463 (size=8101) 2025-01-08T10:47:56,485 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742287_1463 (size=8101) 2025-01-08T10:47:56,485 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:56,491 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b202501088931dc000d18485880dd36fdb85181e2_367c2df29adffa5404639744fd4a5501 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/c4ca4238a0b923820dcc509a6f75849b202501088931dc000d18485880dd36fdb85181e2_367c2df29adffa5404639744fd4a5501 2025-01-08T10:47:56,492 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/367c2df29adffa5404639744fd4a5501/.tmp/cf/48d52513dcf941a0a118141d04e1737c, store: [table=testtb-testExportWithChecksum family=cf region=367c2df29adffa5404639744fd4a5501] 2025-01-08T10:47:56,492 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/367c2df29adffa5404639744fd4a5501/.tmp/cf/48d52513dcf941a0a118141d04e1737c is 206, key is 1db849430b7f3929f6c035c27cdee4247/cf:q/1736333276192/Put/seqid=0 2025-01-08T10:47:56,509 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742288_1464 (size=6108) 2025-01-08T10:47:56,509 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742288_1464 (size=6108) 2025-01-08T10:47:56,509 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742288_1464 (size=6108) 2025-01-08T10:47:56,510 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=266, hasBloomFilter=true, into tmp file hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/6dd967531e3eb9fd666bb33f76786e5c/.tmp/cf/746ed8a4c2a642458956c576a74fbbbe 2025-01-08T10:47:56,515 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/6dd967531e3eb9fd666bb33f76786e5c/.tmp/cf/746ed8a4c2a642458956c576a74fbbbe as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/6dd967531e3eb9fd666bb33f76786e5c/cf/746ed8a4c2a642458956c576a74fbbbe 2025-01-08T10:47:56,519 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/6dd967531e3eb9fd666bb33f76786e5c/cf/746ed8a4c2a642458956c576a74fbbbe, entries=4, sequenceid=6, filesize=6.0 K 2025-01-08T10:47:56,520 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.HRegion(3040): Finished flush of dataSize ~266 B/266, heapSize ~816 B/816, currentSize=0 B/0 for 6dd967531e3eb9fd666bb33f76786e5c in 89ms, sequenceid=6, compaction requested=false 2025-01-08T10:47:56,520 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportWithChecksum' 2025-01-08T10:47:56,520 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.HRegion(2538): Flush status journal for 6dd967531e3eb9fd666bb33f76786e5c: 2025-01-08T10:47:56,520 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c. for snaptb0-testExportWithChecksum completed. 2025-01-08T10:47:56,521 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c.' region-info for snapshot=snaptb0-testExportWithChecksum 2025-01-08T10:47:56,521 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:47:56,521 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/6dd967531e3eb9fd666bb33f76786e5c/cf/746ed8a4c2a642458956c576a74fbbbe] hfiles 2025-01-08T10:47:56,521 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/6dd967531e3eb9fd666bb33f76786e5c/cf/746ed8a4c2a642458956c576a74fbbbe for snapshot=snaptb0-testExportWithChecksum 2025-01-08T10:47:56,530 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742289_1465 (size=14651) 2025-01-08T10:47:56,531 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742289_1465 (size=14651) 2025-01-08T10:47:56,531 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742289_1465 (size=14651) 2025-01-08T10:47:56,532 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=3.0 K, hasBloomFilter=true, into tmp file hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/367c2df29adffa5404639744fd4a5501/.tmp/cf/48d52513dcf941a0a118141d04e1737c 2025-01-08T10:47:56,536 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/367c2df29adffa5404639744fd4a5501/.tmp/cf/48d52513dcf941a0a118141d04e1737c as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/367c2df29adffa5404639744fd4a5501/cf/48d52513dcf941a0a118141d04e1737c 2025-01-08T10:47:56,541 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/367c2df29adffa5404639744fd4a5501/cf/48d52513dcf941a0a118141d04e1737c, entries=46, sequenceid=6, filesize=14.3 K 2025-01-08T10:47:56,542 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.HRegion(3040): Finished flush of dataSize ~3.00 KB/3070, heapSize ~6.70 KB/6864, currentSize=0 B/0 for 367c2df29adffa5404639744fd4a5501 in 111ms, sequenceid=6, compaction requested=false 2025-01-08T10:47:56,542 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.HRegion(2538): Flush status journal for 367c2df29adffa5404639744fd4a5501: 2025-01-08T10:47:56,542 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501. for snaptb0-testExportWithChecksum completed. 2025-01-08T10:47:56,542 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501.' region-info for snapshot=snaptb0-testExportWithChecksum 2025-01-08T10:47:56,542 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:47:56,542 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/367c2df29adffa5404639744fd4a5501/cf/48d52513dcf941a0a118141d04e1737c] hfiles 2025-01-08T10:47:56,542 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/367c2df29adffa5404639744fd4a5501/cf/48d52513dcf941a0a118141d04e1737c for snapshot=snaptb0-testExportWithChecksum 2025-01-08T10:47:56,545 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=188 2025-01-08T10:47:56,560 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742290_1466 (size=107) 2025-01-08T10:47:56,560 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742290_1466 (size=107) 2025-01-08T10:47:56,561 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742290_1466 (size=107) 2025-01-08T10:47:56,561 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c. 2025-01-08T10:47:56,561 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=189 2025-01-08T10:47:56,562 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=189 2025-01-08T10:47:56,562 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportWithChecksum on region 6dd967531e3eb9fd666bb33f76786e5c 2025-01-08T10:47:56,562 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=189, ppid=188, state=RUNNABLE; SnapshotRegionProcedure 6dd967531e3eb9fd666bb33f76786e5c 2025-01-08T10:47:56,565 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=189, ppid=188, state=SUCCESS; SnapshotRegionProcedure 6dd967531e3eb9fd666bb33f76786e5c in 285 msec 2025-01-08T10:47:56,576 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742291_1467 (size=107) 2025-01-08T10:47:56,577 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742291_1467 (size=107) 2025-01-08T10:47:56,577 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742291_1467 (size=107) 2025-01-08T10:47:56,578 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501. 2025-01-08T10:47:56,578 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=190 2025-01-08T10:47:56,578 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=190 2025-01-08T10:47:56,578 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportWithChecksum on region 367c2df29adffa5404639744fd4a5501 2025-01-08T10:47:56,578 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=190, ppid=188, state=RUNNABLE; SnapshotRegionProcedure 367c2df29adffa5404639744fd4a5501 2025-01-08T10:47:56,582 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=190, resume processing ppid=188 2025-01-08T10:47:56,582 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=190, ppid=188, state=SUCCESS; SnapshotRegionProcedure 367c2df29adffa5404639744fd4a5501 in 302 msec 2025-01-08T10:47:56,582 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T10:47:56,583 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T10:47:56,584 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T10:47:56,584 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T10:47:56,584 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:47:56,585 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/c4ca4238a0b923820dcc509a6f75849b202501088931dc000d18485880dd36fdb85181e2_367c2df29adffa5404639744fd4a5501, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/d41d8cd98f00b204e9800998ecf8427e20250108eabde8fd0b394cdaa85e862d9ac36216_6dd967531e3eb9fd666bb33f76786e5c] hfiles 2025-01-08T10:47:56,585 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/c4ca4238a0b923820dcc509a6f75849b202501088931dc000d18485880dd36fdb85181e2_367c2df29adffa5404639744fd4a5501 2025-01-08T10:47:56,585 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/d41d8cd98f00b204e9800998ecf8427e20250108eabde8fd0b394cdaa85e862d9ac36216_6dd967531e3eb9fd666bb33f76786e5c 2025-01-08T10:47:56,591 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742292_1468 (size=291) 2025-01-08T10:47:56,591 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742292_1468 (size=291) 2025-01-08T10:47:56,591 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742292_1468 (size=291) 2025-01-08T10:47:56,592 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T10:47:56,592 DEBUG [PEWorker-4 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testExportWithChecksum 2025-01-08T10:47:56,593 DEBUG [PEWorker-4 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/snaptb0-testExportWithChecksum 2025-01-08T10:47:56,600 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742293_1469 (size=951) 2025-01-08T10:47:56,600 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742293_1469 (size=951) 2025-01-08T10:47:56,600 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742293_1469 (size=951) 2025-01-08T10:47:56,602 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T10:47:56,607 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T10:47:56,607 DEBUG [PEWorker-4 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/snaptb0-testExportWithChecksum to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportWithChecksum 2025-01-08T10:47:56,608 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T10:47:56,608 DEBUG [PEWorker-4 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 }, snapshot procedure id = 188 2025-01-08T10:47:56,609 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=188, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } in 367 msec 2025-01-08T10:47:56,846 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=188 2025-01-08T10:47:56,846 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportWithChecksum, procId: 188 completed 2025-01-08T10:47:56,846 INFO [Time-limited test {}] snapshot.TestExportSnapshot(476): Local export destination path: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333276846 2025-01-08T10:47:56,846 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=file:///, tgtDir=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333276846, rawTgtDir=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333276846, srcFsUri=hdfs://localhost:37495, srcDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:47:56,884 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:37495, inputRoot=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:47:56,884 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=org.apache.hadoop.fs.LocalFileSystem@5d4621f9, outputRoot=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333276846, skipTmp=false, initialOutputSnapshotDir=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333276846/.hbase-snapshot/.tmp/snaptb0-testExportWithChecksum 2025-01-08T10:47:56,885 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T10:47:56,890 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportWithChecksum to file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333276846/.hbase-snapshot/.tmp/snaptb0-testExportWithChecksum 2025-01-08T10:47:57,103 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-10372148436924161493.jar 2025-01-08T10:47:57,104 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:57,104 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:57,104 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:57,595 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportWithChecksum 2025-01-08T10:47:57,596 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportWithChecksum Metrics about Tables on a single HBase RegionServer 2025-01-08T10:47:57,596 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testEmptyExportFileSystemState 2025-01-08T10:47:58,154 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-16313291790166361737.jar 2025-01-08T10:47:58,155 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:58,155 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:58,225 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-7924860958501701982.jar 2025-01-08T10:47:58,225 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:58,225 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:58,225 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:58,226 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:58,226 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:58,226 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T10:47:58,226 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T10:47:58,227 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T10:47:58,227 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T10:47:58,227 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T10:47:58,227 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T10:47:58,228 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T10:47:58,228 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T10:47:58,228 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T10:47:58,228 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T10:47:58,228 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T10:47:58,229 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T10:47:58,229 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T10:47:58,229 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:47:58,229 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:47:58,230 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:47:58,230 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:47:58,230 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:47:58,230 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:47:58,230 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:47:58,287 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742294_1470 (size=127628) 2025-01-08T10:47:58,287 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742294_1470 (size=127628) 2025-01-08T10:47:58,287 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742294_1470 (size=127628) 2025-01-08T10:47:58,297 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742295_1471 (size=2172137) 2025-01-08T10:47:58,298 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742295_1471 (size=2172137) 2025-01-08T10:47:58,298 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742295_1471 (size=2172137) 2025-01-08T10:47:58,308 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742296_1472 (size=213228) 2025-01-08T10:47:58,308 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742296_1472 (size=213228) 2025-01-08T10:47:58,309 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742296_1472 (size=213228) 2025-01-08T10:47:58,324 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742297_1473 (size=1877034) 2025-01-08T10:47:58,324 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742297_1473 (size=1877034) 2025-01-08T10:47:58,324 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742297_1473 (size=1877034) 2025-01-08T10:47:58,334 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742298_1474 (size=533455) 2025-01-08T10:47:58,334 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742298_1474 (size=533455) 2025-01-08T10:47:58,335 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742298_1474 (size=533455) 2025-01-08T10:47:58,358 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742299_1475 (size=7280644) 2025-01-08T10:47:58,358 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742299_1475 (size=7280644) 2025-01-08T10:47:58,358 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742299_1475 (size=7280644) 2025-01-08T10:47:58,381 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742300_1476 (size=6350714) 2025-01-08T10:47:58,381 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742300_1476 (size=6350714) 2025-01-08T10:47:58,382 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742300_1476 (size=6350714) 2025-01-08T10:47:58,401 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742301_1477 (size=4188619) 2025-01-08T10:47:58,402 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742301_1477 (size=4188619) 2025-01-08T10:47:58,402 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742301_1477 (size=4188619) 2025-01-08T10:47:58,408 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742302_1478 (size=20406) 2025-01-08T10:47:58,409 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742302_1478 (size=20406) 2025-01-08T10:47:58,409 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742302_1478 (size=20406) 2025-01-08T10:47:58,415 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742303_1479 (size=75495) 2025-01-08T10:47:58,415 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742303_1479 (size=75495) 2025-01-08T10:47:58,416 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742303_1479 (size=75495) 2025-01-08T10:47:58,422 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742304_1480 (size=45609) 2025-01-08T10:47:58,422 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742304_1480 (size=45609) 2025-01-08T10:47:58,422 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742304_1480 (size=45609) 2025-01-08T10:47:58,429 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742305_1481 (size=110084) 2025-01-08T10:47:58,429 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742305_1481 (size=110084) 2025-01-08T10:47:58,430 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742305_1481 (size=110084) 2025-01-08T10:47:58,441 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742306_1482 (size=1323991) 2025-01-08T10:47:58,441 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742306_1482 (size=1323991) 2025-01-08T10:47:58,442 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742306_1482 (size=1323991) 2025-01-08T10:47:58,448 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742307_1483 (size=23076) 2025-01-08T10:47:58,448 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742307_1483 (size=23076) 2025-01-08T10:47:58,448 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742307_1483 (size=23076) 2025-01-08T10:47:58,454 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742308_1484 (size=126803) 2025-01-08T10:47:58,455 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742308_1484 (size=126803) 2025-01-08T10:47:58,455 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742308_1484 (size=126803) 2025-01-08T10:47:58,470 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742309_1485 (size=322274) 2025-01-08T10:47:58,470 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742309_1485 (size=322274) 2025-01-08T10:47:58,470 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742309_1485 (size=322274) 2025-01-08T10:47:58,479 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742310_1486 (size=912094) 2025-01-08T10:47:58,479 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742310_1486 (size=912094) 2025-01-08T10:47:58,480 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742310_1486 (size=912094) 2025-01-08T10:47:58,491 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742311_1487 (size=1832290) 2025-01-08T10:47:58,491 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742311_1487 (size=1832290) 2025-01-08T10:47:58,491 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742311_1487 (size=1832290) 2025-01-08T10:47:58,498 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742312_1488 (size=30081) 2025-01-08T10:47:58,498 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742312_1488 (size=30081) 2025-01-08T10:47:58,499 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742312_1488 (size=30081) 2025-01-08T10:47:58,516 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742313_1489 (size=53616) 2025-01-08T10:47:58,516 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742313_1489 (size=53616) 2025-01-08T10:47:58,517 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742313_1489 (size=53616) 2025-01-08T10:47:58,535 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742314_1490 (size=29229) 2025-01-08T10:47:58,535 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742314_1490 (size=29229) 2025-01-08T10:47:58,539 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742314_1490 (size=29229) 2025-01-08T10:47:58,551 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742315_1491 (size=169089) 2025-01-08T10:47:58,551 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742315_1491 (size=169089) 2025-01-08T10:47:58,551 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742315_1491 (size=169089) 2025-01-08T10:47:58,576 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742316_1492 (size=5175431) 2025-01-08T10:47:58,576 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742316_1492 (size=5175431) 2025-01-08T10:47:58,576 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742316_1492 (size=5175431) 2025-01-08T10:47:58,585 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742317_1493 (size=451756) 2025-01-08T10:47:58,585 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742317_1493 (size=451756) 2025-01-08T10:47:58,585 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742317_1493 (size=451756) 2025-01-08T10:47:58,592 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742318_1494 (size=136454) 2025-01-08T10:47:58,592 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742318_1494 (size=136454) 2025-01-08T10:47:58,592 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742318_1494 (size=136454) 2025-01-08T10:47:58,607 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742319_1495 (size=3317408) 2025-01-08T10:47:58,607 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742319_1495 (size=3317408) 2025-01-08T10:47:58,607 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742319_1495 (size=3317408) 2025-01-08T10:47:58,615 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742320_1496 (size=503880) 2025-01-08T10:47:58,615 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742320_1496 (size=503880) 2025-01-08T10:47:58,615 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742320_1496 (size=503880) 2025-01-08T10:47:58,636 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742321_1497 (size=4695811) 2025-01-08T10:47:58,637 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742321_1497 (size=4695811) 2025-01-08T10:47:58,637 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742321_1497 (size=4695811) 2025-01-08T10:47:58,638 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T10:47:58,640 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'snaptb0-testExportWithChecksum' hfile list 2025-01-08T10:47:58,641 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(751): export split=0 size=33.2 K 2025-01-08T10:47:58,647 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742322_1498 (size=714) 2025-01-08T10:47:58,648 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742322_1498 (size=714) 2025-01-08T10:47:58,648 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742322_1498 (size=714) 2025-01-08T10:47:58,655 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742323_1499 (size=15) 2025-01-08T10:47:58,655 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742323_1499 (size=15) 2025-01-08T10:47:58,655 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742323_1499 (size=15) 2025-01-08T10:47:58,676 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742324_1500 (size=304977) 2025-01-08T10:47:58,676 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742324_1500 (size=304977) 2025-01-08T10:47:58,676 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742324_1500 (size=304977) 2025-01-08T10:47:59,697 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T10:47:59,697 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T10:47:59,702 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0007_000001 (auth:SIMPLE) from 127.0.0.1:38994 2025-01-08T10:47:59,713 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0008_000001 (auth:SIMPLE) from 127.0.0.1:43494 2025-01-08T10:47:59,715 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_3/usercache/jenkins/appcache/application_1736333116683_0007/container_1736333116683_0007_01_000001/launch_container.sh] 2025-01-08T10:47:59,715 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_3/usercache/jenkins/appcache/application_1736333116683_0007/container_1736333116683_0007_01_000001/container_tokens] 2025-01-08T10:47:59,715 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_3/usercache/jenkins/appcache/application_1736333116683_0007/container_1736333116683_0007_01_000001/sysfs] 2025-01-08T10:48:00,457 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T10:48:05,139 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0008_000001 (auth:SIMPLE) from 127.0.0.1:60532 2025-01-08T10:48:05,380 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742325_1501 (size=350651) 2025-01-08T10:48:05,381 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742325_1501 (size=350651) 2025-01-08T10:48:05,381 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742325_1501 (size=350651) 2025-01-08T10:48:05,714 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T10:48:07,412 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0008_000001 (auth:SIMPLE) from 127.0.0.1:52190 2025-01-08T10:48:12,282 DEBUG [master/67f42ee69b4e:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region a8e8f2d8192bce47f7fedf056a7a63d9 changed from -1.0 to 0.0, refreshing cache 2025-01-08T10:48:12,282 DEBUG [master/67f42ee69b4e:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 6dd967531e3eb9fd666bb33f76786e5c changed from -1.0 to 0.0, refreshing cache 2025-01-08T10:48:12,282 DEBUG [master/67f42ee69b4e:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 367c2df29adffa5404639744fd4a5501 changed from -1.0 to 0.0, refreshing cache 2025-01-08T10:48:12,282 DEBUG [master/67f42ee69b4e:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 5f4e73fdd950eac2bfc23f9cc1241d47 changed from -1.0 to 0.0, refreshing cache 2025-01-08T10:48:12,434 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_0/usercache/jenkins/appcache/application_1736333116683_0008/container_1736333116683_0008_01_000002/launch_container.sh] 2025-01-08T10:48:12,434 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_0/usercache/jenkins/appcache/application_1736333116683_0008/container_1736333116683_0008_01_000002/container_tokens] 2025-01-08T10:48:12,434 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_0/usercache/jenkins/appcache/application_1736333116683_0008/container_1736333116683_0008_01_000002/sysfs] Error: java.io.IOException: Checksum mismatch between hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/367c2df29adffa5404639744fd4a5501/cf/48d52513dcf941a0a118141d04e1737c and file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333276846/archive/data/default/testtb-testExportWithChecksum/367c2df29adffa5404639744fd4a5501/cf/48d52513dcf941a0a118141d04e1737c. Input and output filesystems are of different types. Their checksum algorithms may be incompatible. You can choose file-level checksum validation via -Ddfs.checksum.combine.mode=COMPOSITE_CRC when block-sizes or filesystems are different. Or you can skip checksum-checks altogether with -no-checksum-verify, for the table backup scenario, you should use -i option to skip checksum-checks. (NOTE: By skipping checksums, one runs the risk of masking data-corruption during file-transfer.) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.verifyCopyResult(ExportSnapshot.java:596) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.copyFile(ExportSnapshot.java:332) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.map(ExportSnapshot.java:254) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.map(ExportSnapshot.java:180) at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:145) at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:800) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:348) at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:178) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:172) 2025-01-08T10:48:14,245 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0008_000001 (auth:SIMPLE) from 127.0.0.1:52198 2025-01-08T10:48:17,328 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_2/usercache/jenkins/appcache/application_1736333116683_0008/container_1736333116683_0008_01_000003/launch_container.sh] 2025-01-08T10:48:17,328 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_2/usercache/jenkins/appcache/application_1736333116683_0008/container_1736333116683_0008_01_000003/container_tokens] 2025-01-08T10:48:17,328 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_2/usercache/jenkins/appcache/application_1736333116683_0008/container_1736333116683_0008_01_000003/sysfs] Error: java.io.IOException: Checksum mismatch between hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/367c2df29adffa5404639744fd4a5501/cf/48d52513dcf941a0a118141d04e1737c and file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333276846/archive/data/default/testtb-testExportWithChecksum/367c2df29adffa5404639744fd4a5501/cf/48d52513dcf941a0a118141d04e1737c. Input and output filesystems are of different types. Their checksum algorithms may be incompatible. You can choose file-level checksum validation via -Ddfs.checksum.combine.mode=COMPOSITE_CRC when block-sizes or filesystems are different. Or you can skip checksum-checks altogether with -no-checksum-verify, for the table backup scenario, you should use -i option to skip checksum-checks. (NOTE: By skipping checksums, one runs the risk of masking data-corruption during file-transfer.) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.verifyCopyResult(ExportSnapshot.java:596) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.copyFile(ExportSnapshot.java:332) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.map(ExportSnapshot.java:254) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.map(ExportSnapshot.java:180) at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:145) at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:800) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:348) at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:178) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:172) 2025-01-08T10:48:18,010 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 5f4e73fdd950eac2bfc23f9cc1241d47, had cached 0 bytes from a total of 14859 2025-01-08T10:48:18,010 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region a8e8f2d8192bce47f7fedf056a7a63d9, had cached 0 bytes from a total of 5688 2025-01-08T10:48:19,257 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0008_000001 (auth:SIMPLE) from 127.0.0.1:47262 2025-01-08T10:48:22,169 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_1/usercache/jenkins/appcache/application_1736333116683_0008/container_1736333116683_0008_01_000004/launch_container.sh] 2025-01-08T10:48:22,169 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_1/usercache/jenkins/appcache/application_1736333116683_0008/container_1736333116683_0008_01_000004/container_tokens] 2025-01-08T10:48:22,169 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_1/usercache/jenkins/appcache/application_1736333116683_0008/container_1736333116683_0008_01_000004/sysfs] Error: java.io.IOException: Checksum mismatch between hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/367c2df29adffa5404639744fd4a5501/cf/48d52513dcf941a0a118141d04e1737c and file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/local-export-1736333276846/archive/data/default/testtb-testExportWithChecksum/367c2df29adffa5404639744fd4a5501/cf/48d52513dcf941a0a118141d04e1737c. Input and output filesystems are of different types. Their checksum algorithms may be incompatible. You can choose file-level checksum validation via -Ddfs.checksum.combine.mode=COMPOSITE_CRC when block-sizes or filesystems are different. Or you can skip checksum-checks altogether with -no-checksum-verify, for the table backup scenario, you should use -i option to skip checksum-checks. (NOTE: By skipping checksums, one runs the risk of masking data-corruption during file-transfer.) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.verifyCopyResult(ExportSnapshot.java:596) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.copyFile(ExportSnapshot.java:332) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.map(ExportSnapshot.java:254) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.map(ExportSnapshot.java:180) at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:145) at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:800) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:348) at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:178) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:172) 2025-01-08T10:48:23,268 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0008_000001 (auth:SIMPLE) from 127.0.0.1:42578 2025-01-08T10:48:26,091 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742326_1502 (size=21340) 2025-01-08T10:48:26,091 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742326_1502 (size=21340) 2025-01-08T10:48:26,094 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742326_1502 (size=21340) 2025-01-08T10:48:26,107 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742327_1503 (size=460) 2025-01-08T10:48:26,107 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742327_1503 (size=460) 2025-01-08T10:48:26,108 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742327_1503 (size=460) 2025-01-08T10:48:26,141 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742328_1504 (size=21340) 2025-01-08T10:48:26,142 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742328_1504 (size=21340) 2025-01-08T10:48:26,143 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742328_1504 (size=21340) 2025-01-08T10:48:26,152 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_3/usercache/jenkins/appcache/application_1736333116683_0008/container_1736333116683_0008_01_000005/launch_container.sh] 2025-01-08T10:48:26,152 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_3/usercache/jenkins/appcache/application_1736333116683_0008/container_1736333116683_0008_01_000005/container_tokens] 2025-01-08T10:48:26,152 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_3/usercache/jenkins/appcache/application_1736333116683_0008/container_1736333116683_0008_01_000005/sysfs] 2025-01-08T10:48:26,160 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742329_1505 (size=350651) 2025-01-08T10:48:26,160 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742329_1505 (size=350651) 2025-01-08T10:48:26,161 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742329_1505 (size=350651) 2025-01-08T10:48:26,172 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0008_000001 (auth:SIMPLE) from 127.0.0.1:60366 2025-01-08T10:48:27,920 ERROR [Time-limited test {}] snapshot.ExportSnapshot(1227): Snapshot export failed org.apache.hadoop.hbase.snapshot.ExportSnapshotException: Task failed task_1736333116683_0008_m_000000 Job failed as tasks failed. failedMaps:1 failedReduces:0 killedMaps:0 killedReduces: 0 at org.apache.hadoop.hbase.snapshot.ExportSnapshot.runCopyJob(ExportSnapshot.java:935) ~[classes/:?] at org.apache.hadoop.hbase.snapshot.ExportSnapshot.doWork(ExportSnapshot.java:1204) ~[classes/:?] at org.apache.hadoop.hbase.util.AbstractHBaseTool.run(AbstractHBaseTool.java:151) ~[classes/:?] at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:82) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hbase.snapshot.TestExportSnapshot.runExportSnapshot(TestExportSnapshot.java:523) ~[test-classes/:?] at org.apache.hadoop.hbase.snapshot.TestExportSnapshot.testExportFileSystemState(TestExportSnapshot.java:353) ~[test-classes/:?] at org.apache.hadoop.hbase.snapshot.TestExportSnapshot.testExportWithChecksum(TestExportSnapshot.java:237) ~[test-classes/:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) ~[junit-4.13.2.jar:4.13.2] at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) ~[junit-4.13.2.jar:4.13.2] at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2025-01-08T10:48:27,921 INFO [Time-limited test {}] snapshot.TestExportSnapshot(468): HDFS export destination path: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333307921 2025-01-08T10:48:27,921 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=hdfs://localhost:37495, tgtDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333307921, rawTgtDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333307921, srcFsUri=hdfs://localhost:37495, srcDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:48:27,950 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:37495, inputRoot=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:48:27,950 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-832679797_22, ugi=jenkins (auth:SIMPLE)]], outputRoot=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333307921, skipTmp=false, initialOutputSnapshotDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333307921/.hbase-snapshot/.tmp/snaptb0-testExportWithChecksum 2025-01-08T10:48:27,953 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T10:48:27,959 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportWithChecksum to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333307921/.hbase-snapshot/.tmp/snaptb0-testExportWithChecksum 2025-01-08T10:48:27,973 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742331_1507 (size=156) 2025-01-08T10:48:27,973 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742331_1507 (size=156) 2025-01-08T10:48:27,973 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742331_1507 (size=156) 2025-01-08T10:48:27,979 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742330_1506 (size=951) 2025-01-08T10:48:27,979 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742330_1506 (size=951) 2025-01-08T10:48:27,979 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742330_1506 (size=951) 2025-01-08T10:48:28,151 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-10541212782786685345.jar 2025-01-08T10:48:28,151 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T10:48:28,151 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T10:48:28,152 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T10:48:29,162 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-10898823508255128170.jar 2025-01-08T10:48:29,162 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T10:48:29,163 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T10:48:29,230 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-12639251868210855043.jar 2025-01-08T10:48:29,231 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T10:48:29,231 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T10:48:29,231 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T10:48:29,231 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T10:48:29,231 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T10:48:29,232 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T10:48:29,232 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T10:48:29,232 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T10:48:29,232 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T10:48:29,232 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T10:48:29,233 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T10:48:29,233 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T10:48:29,233 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T10:48:29,233 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T10:48:29,234 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T10:48:29,234 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T10:48:29,234 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T10:48:29,234 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T10:48:29,235 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:48:29,235 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:48:29,235 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:48:29,235 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:48:29,235 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:48:29,236 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:48:29,236 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:48:29,288 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742332_1508 (size=127628) 2025-01-08T10:48:29,288 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742332_1508 (size=127628) 2025-01-08T10:48:29,288 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742332_1508 (size=127628) 2025-01-08T10:48:29,304 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742333_1509 (size=2172137) 2025-01-08T10:48:29,304 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742333_1509 (size=2172137) 2025-01-08T10:48:29,305 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742333_1509 (size=2172137) 2025-01-08T10:48:29,312 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742334_1510 (size=213228) 2025-01-08T10:48:29,312 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742334_1510 (size=213228) 2025-01-08T10:48:29,313 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742334_1510 (size=213228) 2025-01-08T10:48:29,323 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742335_1511 (size=1877034) 2025-01-08T10:48:29,323 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742335_1511 (size=1877034) 2025-01-08T10:48:29,324 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742335_1511 (size=1877034) 2025-01-08T10:48:29,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742336_1512 (size=533455) 2025-01-08T10:48:29,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742336_1512 (size=533455) 2025-01-08T10:48:29,333 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742336_1512 (size=533455) 2025-01-08T10:48:29,359 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742337_1513 (size=7280644) 2025-01-08T10:48:29,359 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742337_1513 (size=7280644) 2025-01-08T10:48:29,360 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742337_1513 (size=7280644) 2025-01-08T10:48:29,381 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742338_1514 (size=4188619) 2025-01-08T10:48:29,381 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742338_1514 (size=4188619) 2025-01-08T10:48:29,382 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742338_1514 (size=4188619) 2025-01-08T10:48:29,388 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742339_1515 (size=20406) 2025-01-08T10:48:29,388 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742339_1515 (size=20406) 2025-01-08T10:48:29,389 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742339_1515 (size=20406) 2025-01-08T10:48:29,394 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742340_1516 (size=75495) 2025-01-08T10:48:29,395 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742340_1516 (size=75495) 2025-01-08T10:48:29,395 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742340_1516 (size=75495) 2025-01-08T10:48:29,403 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742341_1517 (size=451756) 2025-01-08T10:48:29,403 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742341_1517 (size=451756) 2025-01-08T10:48:29,403 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742341_1517 (size=451756) 2025-01-08T10:48:29,410 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742342_1518 (size=45609) 2025-01-08T10:48:29,410 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742342_1518 (size=45609) 2025-01-08T10:48:29,410 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742342_1518 (size=45609) 2025-01-08T10:48:29,420 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742343_1519 (size=110084) 2025-01-08T10:48:29,420 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742343_1519 (size=110084) 2025-01-08T10:48:29,420 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742343_1519 (size=110084) 2025-01-08T10:48:29,430 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742344_1520 (size=1323991) 2025-01-08T10:48:29,430 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742344_1520 (size=1323991) 2025-01-08T10:48:29,430 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742344_1520 (size=1323991) 2025-01-08T10:48:29,437 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742345_1521 (size=23076) 2025-01-08T10:48:29,437 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742345_1521 (size=23076) 2025-01-08T10:48:29,437 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742345_1521 (size=23076) 2025-01-08T10:48:29,443 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742346_1522 (size=126803) 2025-01-08T10:48:29,443 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742346_1522 (size=126803) 2025-01-08T10:48:29,444 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742346_1522 (size=126803) 2025-01-08T10:48:29,450 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742347_1523 (size=322274) 2025-01-08T10:48:29,450 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742347_1523 (size=322274) 2025-01-08T10:48:29,451 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742347_1523 (size=322274) 2025-01-08T10:48:29,469 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742348_1524 (size=1832290) 2025-01-08T10:48:29,469 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742348_1524 (size=1832290) 2025-01-08T10:48:29,469 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742348_1524 (size=1832290) 2025-01-08T10:48:29,477 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742349_1525 (size=30081) 2025-01-08T10:48:29,477 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742349_1525 (size=30081) 2025-01-08T10:48:29,478 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742349_1525 (size=30081) 2025-01-08T10:48:29,491 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742350_1526 (size=53616) 2025-01-08T10:48:29,491 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742350_1526 (size=53616) 2025-01-08T10:48:29,491 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742350_1526 (size=53616) 2025-01-08T10:48:29,502 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742351_1527 (size=29229) 2025-01-08T10:48:29,503 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742351_1527 (size=29229) 2025-01-08T10:48:29,503 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742351_1527 (size=29229) 2025-01-08T10:48:29,510 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742352_1528 (size=169089) 2025-01-08T10:48:29,510 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742352_1528 (size=169089) 2025-01-08T10:48:29,511 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742352_1528 (size=169089) 2025-01-08T10:48:29,520 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742353_1529 (size=912094) 2025-01-08T10:48:29,520 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742353_1529 (size=912094) 2025-01-08T10:48:29,521 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742353_1529 (size=912094) 2025-01-08T10:48:29,546 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742354_1530 (size=5175431) 2025-01-08T10:48:29,546 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742354_1530 (size=5175431) 2025-01-08T10:48:29,546 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742354_1530 (size=5175431) 2025-01-08T10:48:29,552 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742355_1531 (size=136454) 2025-01-08T10:48:29,553 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742355_1531 (size=136454) 2025-01-08T10:48:29,553 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742355_1531 (size=136454) 2025-01-08T10:48:29,571 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742356_1532 (size=6350714) 2025-01-08T10:48:29,572 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742356_1532 (size=6350714) 2025-01-08T10:48:29,572 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742356_1532 (size=6350714) 2025-01-08T10:48:29,585 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742357_1533 (size=3317408) 2025-01-08T10:48:29,586 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742357_1533 (size=3317408) 2025-01-08T10:48:29,586 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742357_1533 (size=3317408) 2025-01-08T10:48:29,593 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742358_1534 (size=503880) 2025-01-08T10:48:29,593 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742358_1534 (size=503880) 2025-01-08T10:48:29,594 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742358_1534 (size=503880) 2025-01-08T10:48:29,613 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742359_1535 (size=4695811) 2025-01-08T10:48:29,613 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742359_1535 (size=4695811) 2025-01-08T10:48:29,613 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742359_1535 (size=4695811) 2025-01-08T10:48:29,614 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T10:48:29,616 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'snaptb0-testExportWithChecksum' hfile list 2025-01-08T10:48:29,618 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(751): export split=0 size=33.2 K 2025-01-08T10:48:29,623 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742360_1536 (size=714) 2025-01-08T10:48:29,623 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742360_1536 (size=714) 2025-01-08T10:48:29,623 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742360_1536 (size=714) 2025-01-08T10:48:29,628 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742361_1537 (size=15) 2025-01-08T10:48:29,628 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742361_1537 (size=15) 2025-01-08T10:48:29,628 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742361_1537 (size=15) 2025-01-08T10:48:29,638 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742362_1538 (size=304929) 2025-01-08T10:48:29,638 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742362_1538 (size=304929) 2025-01-08T10:48:29,639 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742362_1538 (size=304929) 2025-01-08T10:48:32,236 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T10:48:32,236 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T10:48:32,241 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0008_000001 (auth:SIMPLE) from 127.0.0.1:60370 2025-01-08T10:48:32,251 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_0/usercache/jenkins/appcache/application_1736333116683_0008/container_1736333116683_0008_01_000001/launch_container.sh] 2025-01-08T10:48:32,251 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_0/usercache/jenkins/appcache/application_1736333116683_0008/container_1736333116683_0008_01_000001/container_tokens] 2025-01-08T10:48:32,251 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_0/usercache/jenkins/appcache/application_1736333116683_0008/container_1736333116683_0008_01_000001/sysfs] 2025-01-08T10:48:33,162 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0009_000001 (auth:SIMPLE) from 127.0.0.1:35156 2025-01-08T10:48:35,715 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T10:48:37,860 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0009_000001 (auth:SIMPLE) from 127.0.0.1:44474 2025-01-08T10:48:38,073 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742363_1539 (size=350603) 2025-01-08T10:48:38,112 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742363_1539 (size=350603) 2025-01-08T10:48:38,150 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742363_1539 (size=350603) 2025-01-08T10:48:40,109 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0009_000001 (auth:SIMPLE) from 127.0.0.1:58018 2025-01-08T10:48:40,583 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 367c2df29adffa5404639744fd4a5501, had cached 0 bytes from a total of 14651 2025-01-08T10:48:40,583 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 6dd967531e3eb9fd666bb33f76786e5c, had cached 0 bytes from a total of 6108 2025-01-08T10:48:42,903 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742364_1540 (size=14651) 2025-01-08T10:48:42,903 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742364_1540 (size=14651) 2025-01-08T10:48:42,903 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742364_1540 (size=14651) 2025-01-08T10:48:42,967 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742365_1541 (size=8101) 2025-01-08T10:48:42,967 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742365_1541 (size=8101) 2025-01-08T10:48:42,967 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742365_1541 (size=8101) 2025-01-08T10:48:43,017 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742366_1542 (size=6108) 2025-01-08T10:48:43,018 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742366_1542 (size=6108) 2025-01-08T10:48:43,018 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742366_1542 (size=6108) 2025-01-08T10:48:43,061 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742367_1543 (size=5172) 2025-01-08T10:48:43,061 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742367_1543 (size=5172) 2025-01-08T10:48:43,061 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742367_1543 (size=5172) 2025-01-08T10:48:43,140 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742368_1544 (size=17459) 2025-01-08T10:48:43,140 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742368_1544 (size=17459) 2025-01-08T10:48:43,140 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742368_1544 (size=17459) 2025-01-08T10:48:43,156 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742369_1545 (size=462) 2025-01-08T10:48:43,156 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742369_1545 (size=462) 2025-01-08T10:48:43,157 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742369_1545 (size=462) 2025-01-08T10:48:43,189 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742370_1546 (size=17459) 2025-01-08T10:48:43,189 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742370_1546 (size=17459) 2025-01-08T10:48:43,189 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742370_1546 (size=17459) 2025-01-08T10:48:43,200 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_1/usercache/jenkins/appcache/application_1736333116683_0009/container_1736333116683_0009_01_000002/launch_container.sh] 2025-01-08T10:48:43,200 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_1/usercache/jenkins/appcache/application_1736333116683_0009/container_1736333116683_0009_01_000002/container_tokens] 2025-01-08T10:48:43,200 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_1/usercache/jenkins/appcache/application_1736333116683_0009/container_1736333116683_0009_01_000002/sysfs] 2025-01-08T10:48:43,209 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742371_1547 (size=350603) 2025-01-08T10:48:43,209 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742371_1547 (size=350603) 2025-01-08T10:48:43,209 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742371_1547 (size=350603) 2025-01-08T10:48:43,223 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0009_000001 (auth:SIMPLE) from 127.0.0.1:58030 2025-01-08T10:48:45,099 INFO [Time-limited test {}] snapshot.ExportSnapshot(1207): Finalize the Snapshot Export 2025-01-08T10:48:45,100 INFO [Time-limited test {}] snapshot.ExportSnapshot(1218): Verify the exported snapshot's expiration status and integrity. 2025-01-08T10:48:45,106 INFO [Time-limited test {}] snapshot.ExportSnapshot(1224): Export Completed: snaptb0-testExportWithChecksum 2025-01-08T10:48:45,106 INFO [Time-limited test {}] snapshot.TestExportSnapshot(362): Exported snapshot 2025-01-08T10:48:45,106 INFO [Time-limited test {}] snapshot.TestExportSnapshot(373): Verified filesystem state 2025-01-08T10:48:45,106 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-832679797_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportWithChecksum at hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportWithChecksum 2025-01-08T10:48:45,106 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportWithChecksum/.snapshotinfo 2025-01-08T10:48:45,107 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportWithChecksum/data.manifest 2025-01-08T10:48:45,107 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-832679797_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333307921/.hbase-snapshot/snaptb0-testExportWithChecksum at hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333307921/.hbase-snapshot/snaptb0-testExportWithChecksum 2025-01-08T10:48:45,107 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333307921/.hbase-snapshot/snaptb0-testExportWithChecksum/.snapshotinfo 2025-01-08T10:48:45,107 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333307921/.hbase-snapshot/snaptb0-testExportWithChecksum/data.manifest 2025-01-08T10:48:45,112 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testExportWithChecksum 2025-01-08T10:48:45,112 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.2 disable testtb-testExportWithChecksum 2025-01-08T10:48:45,113 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=191, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testExportWithChecksum 2025-01-08T10:48:45,114 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=191 2025-01-08T10:48:45,115 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithChecksum","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333325114"}]},"ts":"1736333325114"} 2025-01-08T10:48:45,116 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithChecksum, state=DISABLING in hbase:meta 2025-01-08T10:48:45,118 INFO [PEWorker-2 {}] procedure.DisableTableProcedure(284): Set testtb-testExportWithChecksum to state=DISABLING 2025-01-08T10:48:45,118 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=192, ppid=191, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testExportWithChecksum}] 2025-01-08T10:48:45,119 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=193, ppid=192, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=6dd967531e3eb9fd666bb33f76786e5c, UNASSIGN}, {pid=194, ppid=192, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=367c2df29adffa5404639744fd4a5501, UNASSIGN}] 2025-01-08T10:48:45,120 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=194, ppid=192, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=367c2df29adffa5404639744fd4a5501, UNASSIGN 2025-01-08T10:48:45,120 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=193, ppid=192, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=6dd967531e3eb9fd666bb33f76786e5c, UNASSIGN 2025-01-08T10:48:45,121 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=194 updating hbase:meta row=367c2df29adffa5404639744fd4a5501, regionState=CLOSING, regionLocation=67f42ee69b4e,38323,1736333108259 2025-01-08T10:48:45,121 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=193 updating hbase:meta row=6dd967531e3eb9fd666bb33f76786e5c, regionState=CLOSING, regionLocation=67f42ee69b4e,36283,1736333108166 2025-01-08T10:48:45,122 DEBUG [PEWorker-1 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T10:48:45,122 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=195, ppid=193, state=RUNNABLE; CloseRegionProcedure 6dd967531e3eb9fd666bb33f76786e5c, server=67f42ee69b4e,36283,1736333108166}] 2025-01-08T10:48:45,122 DEBUG [PEWorker-5 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T10:48:45,122 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=196, ppid=194, state=RUNNABLE; CloseRegionProcedure 367c2df29adffa5404639744fd4a5501, server=67f42ee69b4e,38323,1736333108259}] 2025-01-08T10:48:45,215 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=191 2025-01-08T10:48:45,273 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,36283,1736333108166 2025-01-08T10:48:45,274 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] handler.UnassignRegionHandler(124): Close 6dd967531e3eb9fd666bb33f76786e5c 2025-01-08T10:48:45,274 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T10:48:45,274 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] regionserver.HRegion(1681): Closing 6dd967531e3eb9fd666bb33f76786e5c, disabling compactions & flushes 2025-01-08T10:48:45,274 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] regionserver.HRegion(1703): Closing region testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c. 2025-01-08T10:48:45,274 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c. 2025-01-08T10:48:45,274 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c. after waiting 0 ms 2025-01-08T10:48:45,274 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c. 2025-01-08T10:48:45,274 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,38323,1736333108259 2025-01-08T10:48:45,274 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] handler.UnassignRegionHandler(124): Close 367c2df29adffa5404639744fd4a5501 2025-01-08T10:48:45,275 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T10:48:45,275 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] regionserver.HRegion(1681): Closing 367c2df29adffa5404639744fd4a5501, disabling compactions & flushes 2025-01-08T10:48:45,275 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] regionserver.HRegion(1703): Closing region testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501. 2025-01-08T10:48:45,275 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501. 2025-01-08T10:48:45,275 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501. after waiting 0 ms 2025-01-08T10:48:45,275 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501. 2025-01-08T10:48:45,278 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/6dd967531e3eb9fd666bb33f76786e5c/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T10:48:45,278 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/367c2df29adffa5404639744fd4a5501/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T10:48:45,278 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:48:45,278 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:48:45,278 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] regionserver.HRegion(1922): Closed testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c. 2025-01-08T10:48:45,278 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] regionserver.HRegion(1635): Region close journal for 6dd967531e3eb9fd666bb33f76786e5c: 2025-01-08T10:48:45,278 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] regionserver.HRegion(1922): Closed testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501. 2025-01-08T10:48:45,278 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] regionserver.HRegion(1635): Region close journal for 367c2df29adffa5404639744fd4a5501: 2025-01-08T10:48:45,279 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] handler.UnassignRegionHandler(170): Closed 6dd967531e3eb9fd666bb33f76786e5c 2025-01-08T10:48:45,280 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=193 updating hbase:meta row=6dd967531e3eb9fd666bb33f76786e5c, regionState=CLOSED 2025-01-08T10:48:45,280 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] handler.UnassignRegionHandler(170): Closed 367c2df29adffa5404639744fd4a5501 2025-01-08T10:48:45,280 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=194 updating hbase:meta row=367c2df29adffa5404639744fd4a5501, regionState=CLOSED 2025-01-08T10:48:45,282 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=195, resume processing ppid=193 2025-01-08T10:48:45,283 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=196, resume processing ppid=194 2025-01-08T10:48:45,283 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=196, ppid=194, state=SUCCESS; CloseRegionProcedure 367c2df29adffa5404639744fd4a5501, server=67f42ee69b4e,38323,1736333108259 in 160 msec 2025-01-08T10:48:45,283 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=195, ppid=193, state=SUCCESS; CloseRegionProcedure 6dd967531e3eb9fd666bb33f76786e5c, server=67f42ee69b4e,36283,1736333108166 in 159 msec 2025-01-08T10:48:45,283 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=193, ppid=192, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=6dd967531e3eb9fd666bb33f76786e5c, UNASSIGN in 163 msec 2025-01-08T10:48:45,284 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=194, resume processing ppid=192 2025-01-08T10:48:45,284 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=194, ppid=192, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=367c2df29adffa5404639744fd4a5501, UNASSIGN in 164 msec 2025-01-08T10:48:45,285 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=192, resume processing ppid=191 2025-01-08T10:48:45,286 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=192, ppid=191, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testExportWithChecksum in 166 msec 2025-01-08T10:48:45,286 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithChecksum","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333325286"}]},"ts":"1736333325286"} 2025-01-08T10:48:45,287 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithChecksum, state=DISABLED in hbase:meta 2025-01-08T10:48:45,289 INFO [PEWorker-1 {}] procedure.DisableTableProcedure(296): Set testtb-testExportWithChecksum to state=DISABLED 2025-01-08T10:48:45,290 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=191, state=SUCCESS; DisableTableProcedure table=testtb-testExportWithChecksum in 178 msec 2025-01-08T10:48:45,416 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=191 2025-01-08T10:48:45,416 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testExportWithChecksum, procId: 191 completed 2025-01-08T10:48:45,417 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.2 delete testtb-testExportWithChecksum 2025-01-08T10:48:45,417 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=197, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testExportWithChecksum 2025-01-08T10:48:45,418 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=197, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportWithChecksum 2025-01-08T10:48:45,418 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testExportWithChecksum 2025-01-08T10:48:45,419 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=197, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testExportWithChecksum 2025-01-08T10:48:45,420 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36283 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testExportWithChecksum 2025-01-08T10:48:45,421 DEBUG [HFileArchiver-19 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/6dd967531e3eb9fd666bb33f76786e5c 2025-01-08T10:48:45,421 DEBUG [HFileArchiver-20 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/367c2df29adffa5404639744fd4a5501 2025-01-08T10:48:45,422 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithChecksum 2025-01-08T10:48:45,422 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithChecksum 2025-01-08T10:48:45,422 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithChecksum 2025-01-08T10:48:45,423 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithChecksum 2025-01-08T10:48:45,423 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithChecksum with data PBUF 2025-01-08T10:48:45,423 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithChecksum with data PBUF 2025-01-08T10:48:45,423 DEBUG [HFileArchiver-20 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/367c2df29adffa5404639744fd4a5501/cf, FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/367c2df29adffa5404639744fd4a5501/recovered.edits] 2025-01-08T10:48:45,423 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithChecksum with data PBUF 2025-01-08T10:48:45,423 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithChecksum with data PBUF 2025-01-08T10:48:45,424 DEBUG [HFileArchiver-19 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/6dd967531e3eb9fd666bb33f76786e5c/cf, FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/6dd967531e3eb9fd666bb33f76786e5c/recovered.edits] 2025-01-08T10:48:45,424 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportWithChecksum 2025-01-08T10:48:45,424 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportWithChecksum 2025-01-08T10:48:45,424 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportWithChecksum 2025-01-08T10:48:45,424 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:48:45,424 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:48:45,424 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:48:45,424 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportWithChecksum 2025-01-08T10:48:45,424 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:48:45,425 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=197 2025-01-08T10:48:45,425 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:48:45,425 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:48:45,425 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:48:45,425 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:48:45,427 DEBUG [HFileArchiver-20 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/367c2df29adffa5404639744fd4a5501/cf/48d52513dcf941a0a118141d04e1737c to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportWithChecksum/367c2df29adffa5404639744fd4a5501/cf/48d52513dcf941a0a118141d04e1737c 2025-01-08T10:48:45,428 DEBUG [HFileArchiver-19 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/6dd967531e3eb9fd666bb33f76786e5c/cf/746ed8a4c2a642458956c576a74fbbbe to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportWithChecksum/6dd967531e3eb9fd666bb33f76786e5c/cf/746ed8a4c2a642458956c576a74fbbbe 2025-01-08T10:48:45,430 DEBUG [HFileArchiver-20 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/367c2df29adffa5404639744fd4a5501/recovered.edits/9.seqid to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportWithChecksum/367c2df29adffa5404639744fd4a5501/recovered.edits/9.seqid 2025-01-08T10:48:45,430 DEBUG [HFileArchiver-20 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/367c2df29adffa5404639744fd4a5501 2025-01-08T10:48:45,430 DEBUG [HFileArchiver-19 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/6dd967531e3eb9fd666bb33f76786e5c/recovered.edits/9.seqid to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportWithChecksum/6dd967531e3eb9fd666bb33f76786e5c/recovered.edits/9.seqid 2025-01-08T10:48:45,430 DEBUG [HFileArchiver-19 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportWithChecksum/6dd967531e3eb9fd666bb33f76786e5c 2025-01-08T10:48:45,431 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(313): Archived testtb-testExportWithChecksum regions 2025-01-08T10:48:45,431 DEBUG [PEWorker-3 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6 2025-01-08T10:48:45,431 DEBUG [PEWorker-3 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf] 2025-01-08T10:48:45,434 DEBUG [PEWorker-3 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/c4ca4238a0b923820dcc509a6f75849b202501088931dc000d18485880dd36fdb85181e2_367c2df29adffa5404639744fd4a5501 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/c4ca4238a0b923820dcc509a6f75849b202501088931dc000d18485880dd36fdb85181e2_367c2df29adffa5404639744fd4a5501 2025-01-08T10:48:45,435 DEBUG [PEWorker-3 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/d41d8cd98f00b204e9800998ecf8427e20250108eabde8fd0b394cdaa85e862d9ac36216_6dd967531e3eb9fd666bb33f76786e5c to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/d41d8cd98f00b204e9800998ecf8427e20250108eabde8fd0b394cdaa85e862d9ac36216_6dd967531e3eb9fd666bb33f76786e5c 2025-01-08T10:48:45,435 DEBUG [PEWorker-3 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6 2025-01-08T10:48:45,437 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=197, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testExportWithChecksum 2025-01-08T10:48:45,439 WARN [PEWorker-3 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testtb-testExportWithChecksum from hbase:meta 2025-01-08T10:48:45,440 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testExportWithChecksum' descriptor. 2025-01-08T10:48:45,441 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=197, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testExportWithChecksum 2025-01-08T10:48:45,441 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testExportWithChecksum' from region states. 2025-01-08T10:48:45,441 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736333325441"}]},"ts":"9223372036854775807"} 2025-01-08T10:48:45,441 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736333325441"}]},"ts":"9223372036854775807"} 2025-01-08T10:48:45,443 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T10:48:45,443 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => 6dd967531e3eb9fd666bb33f76786e5c, NAME => 'testtb-testExportWithChecksum,,1736333275229.6dd967531e3eb9fd666bb33f76786e5c.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => 367c2df29adffa5404639744fd4a5501, NAME => 'testtb-testExportWithChecksum,1,1736333275229.367c2df29adffa5404639744fd4a5501.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T10:48:45,443 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testExportWithChecksum' as deleted. 2025-01-08T10:48:45,443 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportWithChecksum","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736333325443"}]},"ts":"9223372036854775807"} 2025-01-08T10:48:45,444 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testExportWithChecksum state from META 2025-01-08T10:48:45,446 DEBUG [PEWorker-3 {}] procedure.DeleteTableProcedure(133): Finished pid=197, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportWithChecksum 2025-01-08T10:48:45,447 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=197, state=SUCCESS; DeleteTableProcedure table=testtb-testExportWithChecksum in 29 msec 2025-01-08T10:48:45,526 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=197 2025-01-08T10:48:45,526 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testExportWithChecksum, procId: 197 completed 2025-01-08T10:48:45,531 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.2 delete name: "emptySnaptb0-testExportWithChecksum" 2025-01-08T10:48:45,533 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(380): Deleting snapshot: emptySnaptb0-testExportWithChecksum 2025-01-08T10:48:45,534 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.2 delete name: "snaptb0-testExportWithChecksum" 2025-01-08T10:48:45,535 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testExportWithChecksum 2025-01-08T10:48:45,612 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: snapshot.TestMobSecureExportSnapshot#testExportWithChecksum Thread=799 (was 806), OpenFileDescriptor=807 (was 827), MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=465 (was 643), ProcessCount=18 (was 18), AvailableMemoryMB=556 (was 1413) 2025-01-08T10:48:45,612 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=799 is superior to 500 2025-01-08T10:48:45,632 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: snapshot.TestMobSecureExportSnapshot#testExportFileSystemStateWithSkipTmp Thread=799, OpenFileDescriptor=807, MaxFileDescriptor=1048576, SystemLoadAverage=465, ProcessCount=18, AvailableMemoryMB=555 2025-01-08T10:48:45,632 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=799 is superior to 500 2025-01-08T10:48:45,633 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.2 create 'testtb-testExportFileSystemStateWithSkipTmp', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T10:48:45,634 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=198, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T10:48:45,635 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=198, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T10:48:45,635 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "testtb-testExportFileSystemStateWithSkipTmp" procId is: 198 2025-01-08T10:48:45,635 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=198, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T10:48:45,635 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=198 2025-01-08T10:48:45,641 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742372_1548 (size=454) 2025-01-08T10:48:45,641 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742372_1548 (size=454) 2025-01-08T10:48:45,641 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742372_1548 (size=454) 2025-01-08T10:48:45,643 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => e98b8450fec8e579da851a8d453681dd, NAME => 'testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testtb-testExportFileSystemStateWithSkipTmp', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:48:45,643 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => b2f5e0d7cfbb791b7590aef92c52e2ba, NAME => 'testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testtb-testExportFileSystemStateWithSkipTmp', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:48:45,649 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742374_1550 (size=79) 2025-01-08T10:48:45,649 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742373_1549 (size=79) 2025-01-08T10:48:45,649 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742374_1550 (size=79) 2025-01-08T10:48:45,650 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742374_1550 (size=79) 2025-01-08T10:48:45,650 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742373_1549 (size=79) 2025-01-08T10:48:45,650 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742373_1549 (size=79) 2025-01-08T10:48:45,650 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:48:45,650 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:48:45,650 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-0 {}] regionserver.HRegion(1681): Closing e98b8450fec8e579da851a8d453681dd, disabling compactions & flushes 2025-01-08T10:48:45,650 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-1 {}] regionserver.HRegion(1681): Closing b2f5e0d7cfbb791b7590aef92c52e2ba, disabling compactions & flushes 2025-01-08T10:48:45,650 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd. 2025-01-08T10:48:45,650 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd. 2025-01-08T10:48:45,650 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd. after waiting 0 ms 2025-01-08T10:48:45,650 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba. 2025-01-08T10:48:45,650 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd. 2025-01-08T10:48:45,650 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba. 2025-01-08T10:48:45,650 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba. after waiting 0 ms 2025-01-08T10:48:45,650 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd. 2025-01-08T10:48:45,650 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba. 2025-01-08T10:48:45,650 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-0 {}] regionserver.HRegion(1635): Region close journal for e98b8450fec8e579da851a8d453681dd: 2025-01-08T10:48:45,650 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba. 2025-01-08T10:48:45,650 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-1 {}] regionserver.HRegion(1635): Region close journal for b2f5e0d7cfbb791b7590aef92c52e2ba: 2025-01-08T10:48:45,651 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=198, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T10:48:45,651 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd.","families":{"info":[{"qualifier":"regioninfo","vlen":78,"tag":[],"timestamp":"1736333325651"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736333325651"}]},"ts":"1736333325651"} 2025-01-08T10:48:45,651 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba.","families":{"info":[{"qualifier":"regioninfo","vlen":78,"tag":[],"timestamp":"1736333325651"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736333325651"}]},"ts":"1736333325651"} 2025-01-08T10:48:45,653 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T10:48:45,654 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=198, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T10:48:45,654 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithSkipTmp","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333325654"}]},"ts":"1736333325654"} 2025-01-08T10:48:45,655 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithSkipTmp, state=ENABLING in hbase:meta 2025-01-08T10:48:45,659 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(202): Hosts are {67f42ee69b4e=0} racks are {/default-rack=0} 2025-01-08T10:48:45,660 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T10:48:45,660 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T10:48:45,660 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T10:48:45,660 INFO [PEWorker-5 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T10:48:45,660 INFO [PEWorker-5 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T10:48:45,660 INFO [PEWorker-5 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T10:48:45,660 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T10:48:45,660 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=199, ppid=198, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=e98b8450fec8e579da851a8d453681dd, ASSIGN}, {pid=200, ppid=198, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=b2f5e0d7cfbb791b7590aef92c52e2ba, ASSIGN}] 2025-01-08T10:48:45,661 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=200, ppid=198, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=b2f5e0d7cfbb791b7590aef92c52e2ba, ASSIGN 2025-01-08T10:48:45,661 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=199, ppid=198, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=e98b8450fec8e579da851a8d453681dd, ASSIGN 2025-01-08T10:48:45,661 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=200, ppid=198, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=b2f5e0d7cfbb791b7590aef92c52e2ba, ASSIGN; state=OFFLINE, location=67f42ee69b4e,38323,1736333108259; forceNewPlan=false, retain=false 2025-01-08T10:48:45,662 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=199, ppid=198, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=e98b8450fec8e579da851a8d453681dd, ASSIGN; state=OFFLINE, location=67f42ee69b4e,45003,1736333108039; forceNewPlan=false, retain=false 2025-01-08T10:48:45,736 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=198 2025-01-08T10:48:45,812 INFO [67f42ee69b4e:34551 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T10:48:45,812 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=199 updating hbase:meta row=e98b8450fec8e579da851a8d453681dd, regionState=OPENING, regionLocation=67f42ee69b4e,45003,1736333108039 2025-01-08T10:48:45,812 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=200 updating hbase:meta row=b2f5e0d7cfbb791b7590aef92c52e2ba, regionState=OPENING, regionLocation=67f42ee69b4e,38323,1736333108259 2025-01-08T10:48:45,814 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=201, ppid=200, state=RUNNABLE; OpenRegionProcedure b2f5e0d7cfbb791b7590aef92c52e2ba, server=67f42ee69b4e,38323,1736333108259}] 2025-01-08T10:48:45,814 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=202, ppid=199, state=RUNNABLE; OpenRegionProcedure e98b8450fec8e579da851a8d453681dd, server=67f42ee69b4e,45003,1736333108039}] 2025-01-08T10:48:45,937 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=198 2025-01-08T10:48:45,965 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,38323,1736333108259 2025-01-08T10:48:45,965 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:48:45,968 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] handler.AssignRegionHandler(135): Open testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba. 2025-01-08T10:48:45,968 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] handler.AssignRegionHandler(135): Open testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd. 2025-01-08T10:48:45,968 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(7285): Opening region: {ENCODED => b2f5e0d7cfbb791b7590aef92c52e2ba, NAME => 'testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T10:48:45,968 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(7285): Opening region: {ENCODED => e98b8450fec8e579da851a8d453681dd, NAME => 'testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T10:48:45,968 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba. service=AccessControlService 2025-01-08T10:48:45,968 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd. service=AccessControlService 2025-01-08T10:48:45,968 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:48:45,968 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T10:48:45,969 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportFileSystemStateWithSkipTmp b2f5e0d7cfbb791b7590aef92c52e2ba 2025-01-08T10:48:45,969 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportFileSystemStateWithSkipTmp e98b8450fec8e579da851a8d453681dd 2025-01-08T10:48:45,969 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:48:45,969 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T10:48:45,969 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(7327): checking encryption for e98b8450fec8e579da851a8d453681dd 2025-01-08T10:48:45,969 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(7327): checking encryption for b2f5e0d7cfbb791b7590aef92c52e2ba 2025-01-08T10:48:45,969 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(7330): checking classloading for e98b8450fec8e579da851a8d453681dd 2025-01-08T10:48:45,969 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(7330): checking classloading for b2f5e0d7cfbb791b7590aef92c52e2ba 2025-01-08T10:48:45,970 INFO [StoreOpener-e98b8450fec8e579da851a8d453681dd-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region e98b8450fec8e579da851a8d453681dd 2025-01-08T10:48:45,970 INFO [StoreOpener-b2f5e0d7cfbb791b7590aef92c52e2ba-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region b2f5e0d7cfbb791b7590aef92c52e2ba 2025-01-08T10:48:45,971 INFO [StoreOpener-e98b8450fec8e579da851a8d453681dd-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region e98b8450fec8e579da851a8d453681dd columnFamilyName cf 2025-01-08T10:48:45,971 INFO [StoreOpener-b2f5e0d7cfbb791b7590aef92c52e2ba-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region b2f5e0d7cfbb791b7590aef92c52e2ba columnFamilyName cf 2025-01-08T10:48:45,972 DEBUG [StoreOpener-e98b8450fec8e579da851a8d453681dd-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:48:45,972 DEBUG [StoreOpener-b2f5e0d7cfbb791b7590aef92c52e2ba-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:48:45,972 INFO [StoreOpener-e98b8450fec8e579da851a8d453681dd-1 {}] regionserver.HStore(327): Store=e98b8450fec8e579da851a8d453681dd/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:48:45,973 INFO [StoreOpener-b2f5e0d7cfbb791b7590aef92c52e2ba-1 {}] regionserver.HStore(327): Store=b2f5e0d7cfbb791b7590aef92c52e2ba/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T10:48:45,973 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/e98b8450fec8e579da851a8d453681dd 2025-01-08T10:48:45,973 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/b2f5e0d7cfbb791b7590aef92c52e2ba 2025-01-08T10:48:45,973 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/e98b8450fec8e579da851a8d453681dd 2025-01-08T10:48:45,974 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/b2f5e0d7cfbb791b7590aef92c52e2ba 2025-01-08T10:48:45,975 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(1085): writing seq id for e98b8450fec8e579da851a8d453681dd 2025-01-08T10:48:45,975 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(1085): writing seq id for b2f5e0d7cfbb791b7590aef92c52e2ba 2025-01-08T10:48:45,976 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/e98b8450fec8e579da851a8d453681dd/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:48:45,976 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/b2f5e0d7cfbb791b7590aef92c52e2ba/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T10:48:45,977 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(1102): Opened e98b8450fec8e579da851a8d453681dd; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=67552219, jitterRate=0.006606504321098328}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T10:48:45,977 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(1102): Opened b2f5e0d7cfbb791b7590aef92c52e2ba; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=60364594, jitterRate=-0.10049745440483093}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T10:48:45,977 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(1001): Region open journal for e98b8450fec8e579da851a8d453681dd: 2025-01-08T10:48:45,977 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(1001): Region open journal for b2f5e0d7cfbb791b7590aef92c52e2ba: 2025-01-08T10:48:45,978 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba., pid=201, masterSystemTime=1736333325965 2025-01-08T10:48:45,978 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd., pid=202, masterSystemTime=1736333325965 2025-01-08T10:48:45,979 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba. 2025-01-08T10:48:45,979 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] handler.AssignRegionHandler(164): Opened testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba. 2025-01-08T10:48:45,980 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=200 updating hbase:meta row=b2f5e0d7cfbb791b7590aef92c52e2ba, regionState=OPEN, openSeqNum=2, regionLocation=67f42ee69b4e,38323,1736333108259 2025-01-08T10:48:45,980 DEBUG [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd. 2025-01-08T10:48:45,980 INFO [RS_OPEN_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] handler.AssignRegionHandler(164): Opened testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd. 2025-01-08T10:48:45,980 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=199 updating hbase:meta row=e98b8450fec8e579da851a8d453681dd, regionState=OPEN, openSeqNum=2, regionLocation=67f42ee69b4e,45003,1736333108039 2025-01-08T10:48:45,982 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=201, resume processing ppid=200 2025-01-08T10:48:45,982 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=201, ppid=200, state=SUCCESS; OpenRegionProcedure b2f5e0d7cfbb791b7590aef92c52e2ba, server=67f42ee69b4e,38323,1736333108259 in 167 msec 2025-01-08T10:48:45,982 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=202, resume processing ppid=199 2025-01-08T10:48:45,982 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=202, ppid=199, state=SUCCESS; OpenRegionProcedure e98b8450fec8e579da851a8d453681dd, server=67f42ee69b4e,45003,1736333108039 in 167 msec 2025-01-08T10:48:45,983 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=200, ppid=198, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=b2f5e0d7cfbb791b7590aef92c52e2ba, ASSIGN in 322 msec 2025-01-08T10:48:45,984 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=199, resume processing ppid=198 2025-01-08T10:48:45,984 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=199, ppid=198, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=e98b8450fec8e579da851a8d453681dd, ASSIGN in 322 msec 2025-01-08T10:48:45,984 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=198, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T10:48:45,984 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithSkipTmp","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333325984"}]},"ts":"1736333325984"} 2025-01-08T10:48:45,985 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithSkipTmp, state=ENABLED in hbase:meta 2025-01-08T10:48:45,987 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=198, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T10:48:45,987 DEBUG [PEWorker-4 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testExportFileSystemStateWithSkipTmp jenkins: RWXCA 2025-01-08T10:48:45,989 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36283 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemStateWithSkipTmp], kv [jenkins: RWXCA] 2025-01-08T10:48:45,992 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:48:45,992 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:48:45,993 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:48:45,993 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:48:45,995 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:48:45,995 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:48:45,995 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:48:45,995 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:48:45,995 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithSkipTmp with data PBUF\x0AQ\x0A\x07jenkins\x12F\x08\x03"B\x0A6\x0A\x07default\x12+testtb-testExportFileSystemStateWithSkipTmp \x00 \x01 \x02 \x03 \x04 2025-01-08T10:48:45,995 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithSkipTmp with data PBUF\x0AQ\x0A\x07jenkins\x12F\x08\x03"B\x0A6\x0A\x07default\x12+testtb-testExportFileSystemStateWithSkipTmp \x00 \x01 \x02 \x03 \x04 2025-01-08T10:48:45,995 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithSkipTmp with data PBUF\x0AQ\x0A\x07jenkins\x12F\x08\x03"B\x0A6\x0A\x07default\x12+testtb-testExportFileSystemStateWithSkipTmp \x00 \x01 \x02 \x03 \x04 2025-01-08T10:48:45,995 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithSkipTmp with data PBUF\x0AQ\x0A\x07jenkins\x12F\x08\x03"B\x0A6\x0A\x07default\x12+testtb-testExportFileSystemStateWithSkipTmp \x00 \x01 \x02 \x03 \x04 2025-01-08T10:48:45,995 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=198, state=SUCCESS; CreateTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp in 361 msec 2025-01-08T10:48:46,238 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=198 2025-01-08T10:48:46,238 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testExportFileSystemStateWithSkipTmp, procId: 198 completed 2025-01-08T10:48:46,240 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T10:48:46,240 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd. 2025-01-08T10:48:46,241 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T10:48:46,251 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.2 snapshot request for:{ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } 2025-01-08T10:48:46,251 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736333326251 (current time:1736333326251). 2025-01-08T10:48:46,251 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T10:48:46,251 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot emptySnaptb0-testExportFileSystemStateWithSkipTmp VERSION not specified, setting to 2 2025-01-08T10:48:46,251 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T10:48:46,252 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x5e9bb627 to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@1c1dae36 2025-01-08T10:48:46,255 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@ec81bca, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:48:46,256 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:48:46,257 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:54716, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:48:46,258 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x5e9bb627 to 127.0.0.1:52367 2025-01-08T10:48:46,259 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:48:46,259 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x16119280 to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@678ce26a 2025-01-08T10:48:46,262 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2954a31a, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:48:46,264 DEBUG [hconnection-0x7d9cee7a-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:48:46,264 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:54722, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:48:46,267 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x16119280 to 127.0.0.1:52367 2025-01-08T10:48:46,267 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:48:46,267 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemStateWithSkipTmp], kv [jenkins: RWXCA] 2025-01-08T10:48:46,267 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T10:48:46,268 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=203, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } 2025-01-08T10:48:46,268 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 }, snapshot procedure id = 203 2025-01-08T10:48:46,269 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T10:48:46,269 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=203 2025-01-08T10:48:46,269 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T10:48:46,271 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T10:48:46,276 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742375_1551 (size=203) 2025-01-08T10:48:46,276 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742375_1551 (size=203) 2025-01-08T10:48:46,277 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742375_1551 (size=203) 2025-01-08T10:48:46,278 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T10:48:46,278 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=204, ppid=203, state=RUNNABLE; SnapshotRegionProcedure e98b8450fec8e579da851a8d453681dd}, {pid=205, ppid=203, state=RUNNABLE; SnapshotRegionProcedure b2f5e0d7cfbb791b7590aef92c52e2ba}] 2025-01-08T10:48:46,279 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=205, ppid=203, state=RUNNABLE; SnapshotRegionProcedure b2f5e0d7cfbb791b7590aef92c52e2ba 2025-01-08T10:48:46,279 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=204, ppid=203, state=RUNNABLE; SnapshotRegionProcedure e98b8450fec8e579da851a8d453681dd 2025-01-08T10:48:46,370 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=203 2025-01-08T10:48:46,429 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,38323,1736333108259 2025-01-08T10:48:46,429 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:48:46,430 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=38323 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=205 2025-01-08T10:48:46,430 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=45003 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=204 2025-01-08T10:48:46,430 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=205}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba. 2025-01-08T10:48:46,430 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=204}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd. 2025-01-08T10:48:46,431 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=204}] regionserver.HRegion(2538): Flush status journal for e98b8450fec8e579da851a8d453681dd: 2025-01-08T10:48:46,431 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=205}] regionserver.HRegion(2538): Flush status journal for b2f5e0d7cfbb791b7590aef92c52e2ba: 2025-01-08T10:48:46,431 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=204}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd. for emptySnaptb0-testExportFileSystemStateWithSkipTmp completed. 2025-01-08T10:48:46,431 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=205}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba. for emptySnaptb0-testExportFileSystemStateWithSkipTmp completed. 2025-01-08T10:48:46,431 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=205}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba.' region-info for snapshot=emptySnaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T10:48:46,431 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=204}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd.' region-info for snapshot=emptySnaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T10:48:46,431 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=205}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:48:46,431 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=204}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:48:46,431 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=204}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T10:48:46,431 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=205}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T10:48:46,436 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742376_1552 (size=82) 2025-01-08T10:48:46,437 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742377_1553 (size=82) 2025-01-08T10:48:46,437 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742376_1552 (size=82) 2025-01-08T10:48:46,437 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742376_1552 (size=82) 2025-01-08T10:48:46,438 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=205}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba. 2025-01-08T10:48:46,438 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742377_1553 (size=82) 2025-01-08T10:48:46,438 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742377_1553 (size=82) 2025-01-08T10:48:46,438 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=205}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=205 2025-01-08T10:48:46,438 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=204}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd. 2025-01-08T10:48:46,438 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=204}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=204 2025-01-08T10:48:46,438 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=205 2025-01-08T10:48:46,438 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportFileSystemStateWithSkipTmp on region b2f5e0d7cfbb791b7590aef92c52e2ba 2025-01-08T10:48:46,439 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=204 2025-01-08T10:48:46,439 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=205, ppid=203, state=RUNNABLE; SnapshotRegionProcedure b2f5e0d7cfbb791b7590aef92c52e2ba 2025-01-08T10:48:46,439 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportFileSystemStateWithSkipTmp on region e98b8450fec8e579da851a8d453681dd 2025-01-08T10:48:46,439 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=204, ppid=203, state=RUNNABLE; SnapshotRegionProcedure e98b8450fec8e579da851a8d453681dd 2025-01-08T10:48:46,440 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=204, ppid=203, state=SUCCESS; SnapshotRegionProcedure e98b8450fec8e579da851a8d453681dd in 161 msec 2025-01-08T10:48:46,441 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=205, resume processing ppid=203 2025-01-08T10:48:46,441 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=205, ppid=203, state=SUCCESS; SnapshotRegionProcedure b2f5e0d7cfbb791b7590aef92c52e2ba in 161 msec 2025-01-08T10:48:46,441 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T10:48:46,441 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T10:48:46,442 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T10:48:46,442 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T10:48:46,442 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:48:46,443 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(328): No files under family: cf 2025-01-08T10:48:46,447 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742378_1554 (size=74) 2025-01-08T10:48:46,448 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742378_1554 (size=74) 2025-01-08T10:48:46,448 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742378_1554 (size=74) 2025-01-08T10:48:46,451 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T10:48:46,451 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for emptySnaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T10:48:46,452 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/emptySnaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T10:48:46,458 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742379_1555 (size=697) 2025-01-08T10:48:46,459 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742379_1555 (size=697) 2025-01-08T10:48:46,459 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742379_1555 (size=697) 2025-01-08T10:48:46,461 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T10:48:46,464 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T10:48:46,465 DEBUG [PEWorker-1 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/emptySnaptb0-testExportFileSystemStateWithSkipTmp to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/emptySnaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T10:48:46,466 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T10:48:46,466 DEBUG [PEWorker-1 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 }, snapshot procedure id = 203 2025-01-08T10:48:46,466 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=203, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } in 199 msec 2025-01-08T10:48:46,499 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportFileSystemStateWithSkipTmp' 2025-01-08T10:48:46,570 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=203 2025-01-08T10:48:46,571 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportFileSystemStateWithSkipTmp, procId: 203 completed 2025-01-08T10:48:46,573 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=45003 {}] regionserver.HRegion(8254): writing data to region testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T10:48:46,586 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38323 {}] regionserver.HRegion(8254): writing data to region testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T10:48:46,589 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T10:48:46,589 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd. 2025-01-08T10:48:46,590 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T10:48:46,599 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.2 snapshot request for:{ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } 2025-01-08T10:48:46,599 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736333326599 (current time:1736333326599). 2025-01-08T10:48:46,599 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T10:48:46,599 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testExportFileSystemStateWithSkipTmp VERSION not specified, setting to 2 2025-01-08T10:48:46,599 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T10:48:46,600 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x5171e9a4 to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@248f0c33 2025-01-08T10:48:46,603 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@35de8816, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:48:46,604 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:48:46,605 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:54732, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:48:46,606 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x5171e9a4 to 127.0.0.1:52367 2025-01-08T10:48:46,606 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:48:46,606 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x28b7d826 to 127.0.0.1:52367 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@7485a797 2025-01-08T10:48:46,609 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@355a1f29, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T10:48:46,610 DEBUG [hconnection-0x6a4c6bb8-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T10:48:46,611 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:54738, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T10:48:46,613 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x28b7d826 to 127.0.0.1:52367 2025-01-08T10:48:46,613 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:48:46,613 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemStateWithSkipTmp], kv [jenkins: RWXCA] 2025-01-08T10:48:46,614 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T10:48:46,615 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=206, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } 2025-01-08T10:48:46,615 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 }, snapshot procedure id = 206 2025-01-08T10:48:46,615 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T10:48:46,615 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=206 2025-01-08T10:48:46,616 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T10:48:46,617 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T10:48:46,625 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742380_1556 (size=198) 2025-01-08T10:48:46,625 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742380_1556 (size=198) 2025-01-08T10:48:46,626 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742380_1556 (size=198) 2025-01-08T10:48:46,626 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T10:48:46,627 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=207, ppid=206, state=RUNNABLE; SnapshotRegionProcedure e98b8450fec8e579da851a8d453681dd}, {pid=208, ppid=206, state=RUNNABLE; SnapshotRegionProcedure b2f5e0d7cfbb791b7590aef92c52e2ba}] 2025-01-08T10:48:46,627 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=208, ppid=206, state=RUNNABLE; SnapshotRegionProcedure b2f5e0d7cfbb791b7590aef92c52e2ba 2025-01-08T10:48:46,627 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=207, ppid=206, state=RUNNABLE; SnapshotRegionProcedure e98b8450fec8e579da851a8d453681dd 2025-01-08T10:48:46,716 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=206 2025-01-08T10:48:46,778 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,38323,1736333108259 2025-01-08T10:48:46,778 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:48:46,779 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=45003 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=207 2025-01-08T10:48:46,779 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=38323 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=208 2025-01-08T10:48:46,779 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba. 2025-01-08T10:48:46,779 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd. 2025-01-08T10:48:46,779 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.HRegion(2837): Flushing e98b8450fec8e579da851a8d453681dd 1/1 column families, dataSize=132 B heapSize=544 B 2025-01-08T10:48:46,779 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.HRegion(2837): Flushing b2f5e0d7cfbb791b7590aef92c52e2ba 1/1 column families, dataSize=3.13 KB heapSize=7 KB 2025-01-08T10:48:46,797 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e202501087a91e33d1d6a4a3abb5ec70e4a831205_e98b8450fec8e579da851a8d453681dd is 71, key is 0a2f3e02a6b94b9736d807e7b1b42a49/cf:q/1736333326573/Put/seqid=0 2025-01-08T10:48:46,797 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108d6408b293fcd4556afd0a5cbb0d601fe_b2f5e0d7cfbb791b7590aef92c52e2ba is 71, key is 12a8d73ec074da512c8ee337956f08a5/cf:q/1736333326586/Put/seqid=0 2025-01-08T10:48:46,802 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742381_1557 (size=5032) 2025-01-08T10:48:46,802 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742381_1557 (size=5032) 2025-01-08T10:48:46,803 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742382_1558 (size=8242) 2025-01-08T10:48:46,803 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742381_1557 (size=5032) 2025-01-08T10:48:46,803 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742382_1558 (size=8242) 2025-01-08T10:48:46,803 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742382_1558 (size=8242) 2025-01-08T10:48:46,804 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:48:46,804 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:48:46,807 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e202501087a91e33d1d6a4a3abb5ec70e4a831205_e98b8450fec8e579da851a8d453681dd to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/d41d8cd98f00b204e9800998ecf8427e202501087a91e33d1d6a4a3abb5ec70e4a831205_e98b8450fec8e579da851a8d453681dd 2025-01-08T10:48:46,807 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108d6408b293fcd4556afd0a5cbb0d601fe_b2f5e0d7cfbb791b7590aef92c52e2ba to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/c4ca4238a0b923820dcc509a6f75849b20250108d6408b293fcd4556afd0a5cbb0d601fe_b2f5e0d7cfbb791b7590aef92c52e2ba 2025-01-08T10:48:46,808 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/e98b8450fec8e579da851a8d453681dd/.tmp/cf/0d8d905ab1104abc8db75f006044a880, store: [table=testtb-testExportFileSystemStateWithSkipTmp family=cf region=e98b8450fec8e579da851a8d453681dd] 2025-01-08T10:48:46,808 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/b2f5e0d7cfbb791b7590aef92c52e2ba/.tmp/cf/bf223fa5ee0944ceae7efc0390fdbf63, store: [table=testtb-testExportFileSystemStateWithSkipTmp family=cf region=b2f5e0d7cfbb791b7590aef92c52e2ba] 2025-01-08T10:48:46,809 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/b2f5e0d7cfbb791b7590aef92c52e2ba/.tmp/cf/bf223fa5ee0944ceae7efc0390fdbf63 is 220, key is 116aa46380355a44ee94e0e3a8bddbb0c/cf:q/1736333326586/Put/seqid=0 2025-01-08T10:48:46,809 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/e98b8450fec8e579da851a8d453681dd/.tmp/cf/0d8d905ab1104abc8db75f006044a880 is 220, key is 0a5eae2975c7c1bde62f9cb61af6b4e6a/cf:q/1736333326573/Put/seqid=0 2025-01-08T10:48:46,813 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742384_1560 (size=15743) 2025-01-08T10:48:46,814 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742384_1560 (size=15743) 2025-01-08T10:48:46,814 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742383_1559 (size=5742) 2025-01-08T10:48:46,814 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742384_1560 (size=15743) 2025-01-08T10:48:46,814 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742383_1559 (size=5742) 2025-01-08T10:48:46,815 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=3.1 K, hasBloomFilter=true, into tmp file hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/b2f5e0d7cfbb791b7590aef92c52e2ba/.tmp/cf/bf223fa5ee0944ceae7efc0390fdbf63 2025-01-08T10:48:46,818 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/b2f5e0d7cfbb791b7590aef92c52e2ba/.tmp/cf/bf223fa5ee0944ceae7efc0390fdbf63 as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/b2f5e0d7cfbb791b7590aef92c52e2ba/cf/bf223fa5ee0944ceae7efc0390fdbf63 2025-01-08T10:48:46,818 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=132, hasBloomFilter=true, into tmp file hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/e98b8450fec8e579da851a8d453681dd/.tmp/cf/0d8d905ab1104abc8db75f006044a880 2025-01-08T10:48:46,819 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742383_1559 (size=5742) 2025-01-08T10:48:46,823 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/e98b8450fec8e579da851a8d453681dd/.tmp/cf/0d8d905ab1104abc8db75f006044a880 as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/e98b8450fec8e579da851a8d453681dd/cf/0d8d905ab1104abc8db75f006044a880 2025-01-08T10:48:46,823 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/b2f5e0d7cfbb791b7590aef92c52e2ba/cf/bf223fa5ee0944ceae7efc0390fdbf63, entries=48, sequenceid=6, filesize=15.4 K 2025-01-08T10:48:46,824 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.HRegion(3040): Finished flush of dataSize ~3.13 KB/3204, heapSize ~6.98 KB/7152, currentSize=0 B/0 for b2f5e0d7cfbb791b7590aef92c52e2ba in 45ms, sequenceid=6, compaction requested=false 2025-01-08T10:48:46,824 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.HRegion(2538): Flush status journal for b2f5e0d7cfbb791b7590aef92c52e2ba: 2025-01-08T10:48:46,824 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba. for snaptb0-testExportFileSystemStateWithSkipTmp completed. 2025-01-08T10:48:46,824 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba.' region-info for snapshot=snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T10:48:46,824 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:48:46,824 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/b2f5e0d7cfbb791b7590aef92c52e2ba/cf/bf223fa5ee0944ceae7efc0390fdbf63] hfiles 2025-01-08T10:48:46,824 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/b2f5e0d7cfbb791b7590aef92c52e2ba/cf/bf223fa5ee0944ceae7efc0390fdbf63 for snapshot=snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T10:48:46,826 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/e98b8450fec8e579da851a8d453681dd/cf/0d8d905ab1104abc8db75f006044a880, entries=2, sequenceid=6, filesize=5.6 K 2025-01-08T10:48:46,827 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.HRegion(3040): Finished flush of dataSize ~132 B/132, heapSize ~528 B/528, currentSize=0 B/0 for e98b8450fec8e579da851a8d453681dd in 48ms, sequenceid=6, compaction requested=false 2025-01-08T10:48:46,827 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.HRegion(2538): Flush status journal for e98b8450fec8e579da851a8d453681dd: 2025-01-08T10:48:46,827 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd. for snaptb0-testExportFileSystemStateWithSkipTmp completed. 2025-01-08T10:48:46,827 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd.' region-info for snapshot=snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T10:48:46,827 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T10:48:46,827 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/e98b8450fec8e579da851a8d453681dd/cf/0d8d905ab1104abc8db75f006044a880] hfiles 2025-01-08T10:48:46,827 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/e98b8450fec8e579da851a8d453681dd/cf/0d8d905ab1104abc8db75f006044a880 for snapshot=snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T10:48:46,829 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742385_1561 (size=121) 2025-01-08T10:48:46,829 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742385_1561 (size=121) 2025-01-08T10:48:46,830 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742385_1561 (size=121) 2025-01-08T10:48:46,830 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba. 2025-01-08T10:48:46,830 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=208 2025-01-08T10:48:46,831 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=208 2025-01-08T10:48:46,831 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportFileSystemStateWithSkipTmp on region b2f5e0d7cfbb791b7590aef92c52e2ba 2025-01-08T10:48:46,831 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=208, ppid=206, state=RUNNABLE; SnapshotRegionProcedure b2f5e0d7cfbb791b7590aef92c52e2ba 2025-01-08T10:48:46,832 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=208, ppid=206, state=SUCCESS; SnapshotRegionProcedure b2f5e0d7cfbb791b7590aef92c52e2ba in 204 msec 2025-01-08T10:48:46,833 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742386_1562 (size=121) 2025-01-08T10:48:46,833 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742386_1562 (size=121) 2025-01-08T10:48:46,834 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742386_1562 (size=121) 2025-01-08T10:48:46,834 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd. 2025-01-08T10:48:46,834 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/67f42ee69b4e:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=207 2025-01-08T10:48:46,834 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.HMaster(4121): Remote procedure done, pid=207 2025-01-08T10:48:46,834 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportFileSystemStateWithSkipTmp on region e98b8450fec8e579da851a8d453681dd 2025-01-08T10:48:46,835 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=207, ppid=206, state=RUNNABLE; SnapshotRegionProcedure e98b8450fec8e579da851a8d453681dd 2025-01-08T10:48:46,836 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=207, resume processing ppid=206 2025-01-08T10:48:46,836 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=207, ppid=206, state=SUCCESS; SnapshotRegionProcedure e98b8450fec8e579da851a8d453681dd in 208 msec 2025-01-08T10:48:46,836 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T10:48:46,837 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T10:48:46,837 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T10:48:46,837 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T10:48:46,837 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T10:48:46,838 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/c4ca4238a0b923820dcc509a6f75849b20250108d6408b293fcd4556afd0a5cbb0d601fe_b2f5e0d7cfbb791b7590aef92c52e2ba, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/d41d8cd98f00b204e9800998ecf8427e202501087a91e33d1d6a4a3abb5ec70e4a831205_e98b8450fec8e579da851a8d453681dd] hfiles 2025-01-08T10:48:46,838 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/c4ca4238a0b923820dcc509a6f75849b20250108d6408b293fcd4556afd0a5cbb0d601fe_b2f5e0d7cfbb791b7590aef92c52e2ba 2025-01-08T10:48:46,838 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/d41d8cd98f00b204e9800998ecf8427e202501087a91e33d1d6a4a3abb5ec70e4a831205_e98b8450fec8e579da851a8d453681dd 2025-01-08T10:48:46,843 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742387_1563 (size=305) 2025-01-08T10:48:46,843 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742387_1563 (size=305) 2025-01-08T10:48:46,843 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742387_1563 (size=305) 2025-01-08T10:48:46,844 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T10:48:46,844 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T10:48:46,845 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T10:48:46,850 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742388_1564 (size=1007) 2025-01-08T10:48:46,851 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742388_1564 (size=1007) 2025-01-08T10:48:46,851 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742388_1564 (size=1007) 2025-01-08T10:48:46,853 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T10:48:46,857 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T10:48:46,857 DEBUG [PEWorker-3 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemStateWithSkipTmp to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T10:48:46,858 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T10:48:46,858 DEBUG [PEWorker-3 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 }, snapshot procedure id = 206 2025-01-08T10:48:46,859 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=206, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } in 244 msec 2025-01-08T10:48:46,917 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=206 2025-01-08T10:48:46,917 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportFileSystemStateWithSkipTmp, procId: 206 completed 2025-01-08T10:48:46,917 INFO [Time-limited test {}] snapshot.TestExportSnapshot(468): HDFS export destination path: hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333326917 2025-01-08T10:48:46,918 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=hdfs://localhost:37495, tgtDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333326917, rawTgtDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333326917, srcFsUri=hdfs://localhost:37495, srcDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:48:46,955 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:37495, inputRoot=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f 2025-01-08T10:48:46,955 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-832679797_22, ugi=jenkins (auth:SIMPLE)]], outputRoot=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333326917, skipTmp=true, initialOutputSnapshotDir=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333326917/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T10:48:46,956 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T10:48:46,960 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333326917/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T10:48:46,971 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742390_1566 (size=198) 2025-01-08T10:48:46,971 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742390_1566 (size=198) 2025-01-08T10:48:46,971 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742390_1566 (size=198) 2025-01-08T10:48:46,978 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742389_1565 (size=1007) 2025-01-08T10:48:46,978 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742389_1565 (size=1007) 2025-01-08T10:48:46,978 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742389_1565 (size=1007) 2025-01-08T10:48:47,149 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-2956729214867904054.jar 2025-01-08T10:48:47,150 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T10:48:47,150 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T10:48:47,150 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T10:48:47,595 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T10:48:47,596 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemStateWithSkipTmp Metrics about Tables on a single HBase RegionServer 2025-01-08T10:48:47,596 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportWithChecksum 2025-01-08T10:48:48,140 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-170103543471444522.jar 2025-01-08T10:48:48,140 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T10:48:48,141 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T10:48:48,211 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop-8447827956114968006.jar 2025-01-08T10:48:48,212 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T10:48:48,212 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T10:48:48,212 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T10:48:48,213 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T10:48:48,213 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T10:48:48,213 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T10:48:48,214 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T10:48:48,214 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T10:48:48,214 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T10:48:48,214 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T10:48:48,215 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T10:48:48,215 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T10:48:48,215 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T10:48:48,215 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T10:48:48,215 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T10:48:48,216 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T10:48:48,216 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T10:48:48,216 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T10:48:48,217 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:48:48,217 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:48:48,217 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:48:48,217 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:48:48,217 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T10:48:48,218 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:48:48,218 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T10:48:48,269 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742391_1567 (size=127628) 2025-01-08T10:48:48,269 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742391_1567 (size=127628) 2025-01-08T10:48:48,269 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742391_1567 (size=127628) 2025-01-08T10:48:48,278 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742392_1568 (size=2172137) 2025-01-08T10:48:48,278 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742392_1568 (size=2172137) 2025-01-08T10:48:48,279 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742392_1568 (size=2172137) 2025-01-08T10:48:48,284 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742393_1569 (size=213228) 2025-01-08T10:48:48,285 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742393_1569 (size=213228) 2025-01-08T10:48:48,285 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742393_1569 (size=213228) 2025-01-08T10:48:48,294 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742394_1570 (size=1877034) 2025-01-08T10:48:48,295 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742394_1570 (size=1877034) 2025-01-08T10:48:48,295 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742394_1570 (size=1877034) 2025-01-08T10:48:48,302 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742395_1571 (size=533455) 2025-01-08T10:48:48,302 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742395_1571 (size=533455) 2025-01-08T10:48:48,303 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742395_1571 (size=533455) 2025-01-08T10:48:48,325 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742396_1572 (size=7280644) 2025-01-08T10:48:48,325 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742396_1572 (size=7280644) 2025-01-08T10:48:48,326 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742396_1572 (size=7280644) 2025-01-08T10:48:48,341 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742397_1573 (size=4188619) 2025-01-08T10:48:48,341 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742397_1573 (size=4188619) 2025-01-08T10:48:48,342 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742397_1573 (size=4188619) 2025-01-08T10:48:48,364 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742398_1574 (size=6350714) 2025-01-08T10:48:48,365 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742398_1574 (size=6350714) 2025-01-08T10:48:48,365 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742398_1574 (size=6350714) 2025-01-08T10:48:48,371 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742399_1575 (size=20406) 2025-01-08T10:48:48,372 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742399_1575 (size=20406) 2025-01-08T10:48:48,372 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742399_1575 (size=20406) 2025-01-08T10:48:48,378 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742400_1576 (size=75495) 2025-01-08T10:48:48,378 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742400_1576 (size=75495) 2025-01-08T10:48:48,379 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742400_1576 (size=75495) 2025-01-08T10:48:48,384 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742401_1577 (size=45609) 2025-01-08T10:48:48,384 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742401_1577 (size=45609) 2025-01-08T10:48:48,385 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742401_1577 (size=45609) 2025-01-08T10:48:48,390 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742402_1578 (size=110084) 2025-01-08T10:48:48,390 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742402_1578 (size=110084) 2025-01-08T10:48:48,391 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742402_1578 (size=110084) 2025-01-08T10:48:48,399 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742403_1579 (size=1323991) 2025-01-08T10:48:48,399 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742403_1579 (size=1323991) 2025-01-08T10:48:48,399 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742403_1579 (size=1323991) 2025-01-08T10:48:48,404 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742404_1580 (size=23076) 2025-01-08T10:48:48,404 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742404_1580 (size=23076) 2025-01-08T10:48:48,405 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742404_1580 (size=23076) 2025-01-08T10:48:48,410 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742405_1581 (size=126803) 2025-01-08T10:48:48,410 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742405_1581 (size=126803) 2025-01-08T10:48:48,411 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742405_1581 (size=126803) 2025-01-08T10:48:48,416 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742406_1582 (size=322274) 2025-01-08T10:48:48,416 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742406_1582 (size=322274) 2025-01-08T10:48:48,417 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742406_1582 (size=322274) 2025-01-08T10:48:48,426 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742407_1583 (size=1832290) 2025-01-08T10:48:48,426 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742407_1583 (size=1832290) 2025-01-08T10:48:48,426 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742407_1583 (size=1832290) 2025-01-08T10:48:48,432 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742408_1584 (size=30081) 2025-01-08T10:48:48,432 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742408_1584 (size=30081) 2025-01-08T10:48:48,432 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742408_1584 (size=30081) 2025-01-08T10:48:48,437 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742409_1585 (size=53616) 2025-01-08T10:48:48,437 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742409_1585 (size=53616) 2025-01-08T10:48:48,438 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742409_1585 (size=53616) 2025-01-08T10:48:48,443 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742410_1586 (size=29229) 2025-01-08T10:48:48,443 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742410_1586 (size=29229) 2025-01-08T10:48:48,444 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742410_1586 (size=29229) 2025-01-08T10:48:48,449 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742411_1587 (size=169089) 2025-01-08T10:48:48,450 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742411_1587 (size=169089) 2025-01-08T10:48:48,450 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742411_1587 (size=169089) 2025-01-08T10:48:48,456 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742412_1588 (size=451756) 2025-01-08T10:48:48,456 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742412_1588 (size=451756) 2025-01-08T10:48:48,456 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742412_1588 (size=451756) 2025-01-08T10:48:48,475 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742413_1589 (size=5175431) 2025-01-08T10:48:48,476 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742413_1589 (size=5175431) 2025-01-08T10:48:48,476 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742413_1589 (size=5175431) 2025-01-08T10:48:48,482 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742414_1590 (size=136454) 2025-01-08T10:48:48,482 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742414_1590 (size=136454) 2025-01-08T10:48:48,483 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742414_1590 (size=136454) 2025-01-08T10:48:48,491 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742415_1591 (size=912094) 2025-01-08T10:48:48,491 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742415_1591 (size=912094) 2025-01-08T10:48:48,491 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742415_1591 (size=912094) 2025-01-08T10:48:48,503 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742416_1592 (size=3317408) 2025-01-08T10:48:48,503 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742416_1592 (size=3317408) 2025-01-08T10:48:48,504 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742416_1592 (size=3317408) 2025-01-08T10:48:48,511 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742417_1593 (size=503880) 2025-01-08T10:48:48,511 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742417_1593 (size=503880) 2025-01-08T10:48:48,511 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742417_1593 (size=503880) 2025-01-08T10:48:48,527 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742418_1594 (size=4695811) 2025-01-08T10:48:48,527 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742418_1594 (size=4695811) 2025-01-08T10:48:48,527 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742418_1594 (size=4695811) 2025-01-08T10:48:48,528 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T10:48:48,530 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'snaptb0-testExportFileSystemStateWithSkipTmp' hfile list 2025-01-08T10:48:48,531 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(751): export split=0 size=33.9 K 2025-01-08T10:48:48,536 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742419_1595 (size=770) 2025-01-08T10:48:48,536 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742419_1595 (size=770) 2025-01-08T10:48:48,537 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742419_1595 (size=770) 2025-01-08T10:48:48,541 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742420_1596 (size=15) 2025-01-08T10:48:48,541 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742420_1596 (size=15) 2025-01-08T10:48:48,542 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742420_1596 (size=15) 2025-01-08T10:48:48,555 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742421_1597 (size=305095) 2025-01-08T10:48:48,555 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742421_1597 (size=305095) 2025-01-08T10:48:48,555 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742421_1597 (size=305095) 2025-01-08T10:48:49,287 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T10:48:49,287 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T10:48:49,289 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0009_000001 (auth:SIMPLE) from 127.0.0.1:44416 2025-01-08T10:48:49,301 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_3/usercache/jenkins/appcache/application_1736333116683_0009/container_1736333116683_0009_01_000001/launch_container.sh] 2025-01-08T10:48:49,301 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_3/usercache/jenkins/appcache/application_1736333116683_0009/container_1736333116683_0009_01_000001/container_tokens] 2025-01-08T10:48:49,301 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-0_3/usercache/jenkins/appcache/application_1736333116683_0009/container_1736333116683_0009_01_000001/sysfs] 2025-01-08T10:48:50,245 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0010_000001 (auth:SIMPLE) from 127.0.0.1:59328 2025-01-08T10:48:50,810 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T10:48:55,176 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0010_000001 (auth:SIMPLE) from 127.0.0.1:39806 2025-01-08T10:48:55,381 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742422_1598 (size=350793) 2025-01-08T10:48:55,381 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742422_1598 (size=350793) 2025-01-08T10:48:55,381 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742422_1598 (size=350793) 2025-01-08T10:48:57,402 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0010_000001 (auth:SIMPLE) from 127.0.0.1:49468 2025-01-08T10:49:00,111 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742423_1599 (size=15743) 2025-01-08T10:49:00,111 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742423_1599 (size=15743) 2025-01-08T10:49:00,112 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742423_1599 (size=15743) 2025-01-08T10:49:00,166 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742424_1600 (size=8242) 2025-01-08T10:49:00,166 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742424_1600 (size=8242) 2025-01-08T10:49:00,166 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742424_1600 (size=8242) 2025-01-08T10:49:00,207 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742425_1601 (size=5742) 2025-01-08T10:49:00,207 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742425_1601 (size=5742) 2025-01-08T10:49:00,208 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742425_1601 (size=5742) 2025-01-08T10:49:00,247 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742426_1602 (size=5032) 2025-01-08T10:49:00,247 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742426_1602 (size=5032) 2025-01-08T10:49:00,247 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742426_1602 (size=5032) 2025-01-08T10:49:00,330 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742427_1603 (size=17473) 2025-01-08T10:49:00,331 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742427_1603 (size=17473) 2025-01-08T10:49:00,331 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742427_1603 (size=17473) 2025-01-08T10:49:00,342 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742428_1604 (size=476) 2025-01-08T10:49:00,343 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742428_1604 (size=476) 2025-01-08T10:49:00,343 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742428_1604 (size=476) 2025-01-08T10:49:00,371 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742429_1605 (size=17473) 2025-01-08T10:49:00,371 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742429_1605 (size=17473) 2025-01-08T10:49:00,371 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742429_1605 (size=17473) 2025-01-08T10:49:00,379 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_3/usercache/jenkins/appcache/application_1736333116683_0010/container_1736333116683_0010_01_000002/launch_container.sh] 2025-01-08T10:49:00,379 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_3/usercache/jenkins/appcache/application_1736333116683_0010/container_1736333116683_0010_01_000002/container_tokens] 2025-01-08T10:49:00,379 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_3/usercache/jenkins/appcache/application_1736333116683_0010/container_1736333116683_0010_01_000002/sysfs] 2025-01-08T10:49:00,388 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742430_1606 (size=350793) 2025-01-08T10:49:00,388 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742430_1606 (size=350793) 2025-01-08T10:49:00,389 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742430_1606 (size=350793) 2025-01-08T10:49:00,402 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736333116683_0010_000001 (auth:SIMPLE) from 127.0.0.1:49472 2025-01-08T10:49:01,679 INFO [Time-limited test {}] snapshot.ExportSnapshot(1207): Finalize the Snapshot Export 2025-01-08T10:49:01,680 INFO [Time-limited test {}] snapshot.ExportSnapshot(1218): Verify the exported snapshot's expiration status and integrity. 2025-01-08T10:49:01,686 INFO [Time-limited test {}] snapshot.ExportSnapshot(1224): Export Completed: snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T10:49:01,686 INFO [Time-limited test {}] snapshot.TestExportSnapshot(362): Exported snapshot 2025-01-08T10:49:01,687 INFO [Time-limited test {}] snapshot.TestExportSnapshot(373): Verified filesystem state 2025-01-08T10:49:01,687 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-832679797_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp at hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T10:49:01,687 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp/.snapshotinfo 2025-01-08T10:49:01,687 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp/data.manifest 2025-01-08T10:49:01,687 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-832679797_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333326917/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp at hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333326917/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T10:49:01,688 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333326917/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp/.snapshotinfo 2025-01-08T10:49:01,688 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/export-test/export-1736333326917/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp/data.manifest 2025-01-08T10:49:01,693 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T10:49:01,693 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.2 disable testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T10:49:01,694 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=209, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T10:49:01,695 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=209 2025-01-08T10:49:01,695 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithSkipTmp","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333341695"}]},"ts":"1736333341695"} 2025-01-08T10:49:01,696 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithSkipTmp, state=DISABLING in hbase:meta 2025-01-08T10:49:01,698 INFO [PEWorker-2 {}] procedure.DisableTableProcedure(284): Set testtb-testExportFileSystemStateWithSkipTmp to state=DISABLING 2025-01-08T10:49:01,699 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=210, ppid=209, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testExportFileSystemStateWithSkipTmp}] 2025-01-08T10:49:01,700 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=211, ppid=210, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=e98b8450fec8e579da851a8d453681dd, UNASSIGN}, {pid=212, ppid=210, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=b2f5e0d7cfbb791b7590aef92c52e2ba, UNASSIGN}] 2025-01-08T10:49:01,701 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=212, ppid=210, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=b2f5e0d7cfbb791b7590aef92c52e2ba, UNASSIGN 2025-01-08T10:49:01,701 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=211, ppid=210, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=e98b8450fec8e579da851a8d453681dd, UNASSIGN 2025-01-08T10:49:01,702 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=211 updating hbase:meta row=e98b8450fec8e579da851a8d453681dd, regionState=CLOSING, regionLocation=67f42ee69b4e,45003,1736333108039 2025-01-08T10:49:01,702 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=212 updating hbase:meta row=b2f5e0d7cfbb791b7590aef92c52e2ba, regionState=CLOSING, regionLocation=67f42ee69b4e,38323,1736333108259 2025-01-08T10:49:01,703 DEBUG [PEWorker-5 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T10:49:01,703 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=213, ppid=212, state=RUNNABLE; CloseRegionProcedure b2f5e0d7cfbb791b7590aef92c52e2ba, server=67f42ee69b4e,38323,1736333108259}] 2025-01-08T10:49:01,703 DEBUG [PEWorker-1 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T10:49:01,703 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=214, ppid=211, state=RUNNABLE; CloseRegionProcedure e98b8450fec8e579da851a8d453681dd, server=67f42ee69b4e,45003,1736333108039}] 2025-01-08T10:49:01,796 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=209 2025-01-08T10:49:01,854 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,45003,1736333108039 2025-01-08T10:49:01,854 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 67f42ee69b4e,38323,1736333108259 2025-01-08T10:49:01,854 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] handler.UnassignRegionHandler(124): Close e98b8450fec8e579da851a8d453681dd 2025-01-08T10:49:01,854 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] handler.UnassignRegionHandler(124): Close b2f5e0d7cfbb791b7590aef92c52e2ba 2025-01-08T10:49:01,855 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T10:49:01,855 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T10:49:01,855 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] regionserver.HRegion(1681): Closing b2f5e0d7cfbb791b7590aef92c52e2ba, disabling compactions & flushes 2025-01-08T10:49:01,855 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] regionserver.HRegion(1681): Closing e98b8450fec8e579da851a8d453681dd, disabling compactions & flushes 2025-01-08T10:49:01,855 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd. 2025-01-08T10:49:01,855 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba. 2025-01-08T10:49:01,855 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba. 2025-01-08T10:49:01,855 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd. 2025-01-08T10:49:01,855 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba. after waiting 0 ms 2025-01-08T10:49:01,855 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd. after waiting 0 ms 2025-01-08T10:49:01,855 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba. 2025-01-08T10:49:01,855 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd. 2025-01-08T10:49:01,859 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/b2f5e0d7cfbb791b7590aef92c52e2ba/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T10:49:01,859 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/e98b8450fec8e579da851a8d453681dd/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T10:49:01,859 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:49:01,860 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba. 2025-01-08T10:49:01,860 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] regionserver.HRegion(1635): Region close journal for b2f5e0d7cfbb791b7590aef92c52e2ba: 2025-01-08T10:49:01,860 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:49:01,860 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd. 2025-01-08T10:49:01,860 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] regionserver.HRegion(1635): Region close journal for e98b8450fec8e579da851a8d453681dd: 2025-01-08T10:49:01,861 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] handler.UnassignRegionHandler(170): Closed b2f5e0d7cfbb791b7590aef92c52e2ba 2025-01-08T10:49:01,861 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=212 updating hbase:meta row=b2f5e0d7cfbb791b7590aef92c52e2ba, regionState=CLOSED 2025-01-08T10:49:01,861 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] handler.UnassignRegionHandler(170): Closed e98b8450fec8e579da851a8d453681dd 2025-01-08T10:49:01,862 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=211 updating hbase:meta row=e98b8450fec8e579da851a8d453681dd, regionState=CLOSED 2025-01-08T10:49:01,864 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=213, resume processing ppid=212 2025-01-08T10:49:01,864 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=213, ppid=212, state=SUCCESS; CloseRegionProcedure b2f5e0d7cfbb791b7590aef92c52e2ba, server=67f42ee69b4e,38323,1736333108259 in 159 msec 2025-01-08T10:49:01,864 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=214, resume processing ppid=211 2025-01-08T10:49:01,864 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=214, ppid=211, state=SUCCESS; CloseRegionProcedure e98b8450fec8e579da851a8d453681dd, server=67f42ee69b4e,45003,1736333108039 in 160 msec 2025-01-08T10:49:01,864 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=212, ppid=210, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=b2f5e0d7cfbb791b7590aef92c52e2ba, UNASSIGN in 164 msec 2025-01-08T10:49:01,865 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=211, resume processing ppid=210 2025-01-08T10:49:01,865 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=211, ppid=210, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=e98b8450fec8e579da851a8d453681dd, UNASSIGN in 164 msec 2025-01-08T10:49:01,867 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=210, resume processing ppid=209 2025-01-08T10:49:01,867 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=210, ppid=209, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testExportFileSystemStateWithSkipTmp in 167 msec 2025-01-08T10:49:01,867 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithSkipTmp","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736333341867"}]},"ts":"1736333341867"} 2025-01-08T10:49:01,868 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithSkipTmp, state=DISABLED in hbase:meta 2025-01-08T10:49:01,870 INFO [PEWorker-4 {}] procedure.DisableTableProcedure(296): Set testtb-testExportFileSystemStateWithSkipTmp to state=DISABLED 2025-01-08T10:49:01,871 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=209, state=SUCCESS; DisableTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp in 177 msec 2025-01-08T10:49:01,996 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=209 2025-01-08T10:49:01,996 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testExportFileSystemStateWithSkipTmp, procId: 209 completed 2025-01-08T10:49:01,997 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.2 delete testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T10:49:01,997 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] procedure2.ProcedureExecutor(1098): Stored pid=215, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T10:49:01,998 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=215, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T10:49:01,998 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T10:49:01,999 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=215, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T10:49:02,000 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36283 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T10:49:02,001 DEBUG [HFileArchiver-21 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/e98b8450fec8e579da851a8d453681dd 2025-01-08T10:49:02,002 DEBUG [HFileArchiver-22 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/b2f5e0d7cfbb791b7590aef92c52e2ba 2025-01-08T10:49:02,003 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T10:49:02,003 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T10:49:02,003 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T10:49:02,003 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T10:49:02,003 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithSkipTmp with data PBUF 2025-01-08T10:49:02,003 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithSkipTmp with data PBUF 2025-01-08T10:49:02,004 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithSkipTmp with data PBUF 2025-01-08T10:49:02,004 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithSkipTmp with data PBUF 2025-01-08T10:49:02,004 DEBUG [HFileArchiver-22 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/b2f5e0d7cfbb791b7590aef92c52e2ba/cf, FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/b2f5e0d7cfbb791b7590aef92c52e2ba/recovered.edits] 2025-01-08T10:49:02,004 DEBUG [HFileArchiver-21 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/e98b8450fec8e579da851a8d453681dd/cf, FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/e98b8450fec8e579da851a8d453681dd/recovered.edits] 2025-01-08T10:49:02,005 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T10:49:02,005 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T10:49:02,005 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:49:02,005 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:49:02,005 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T10:49:02,005 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:49:02,005 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T10:49:02,005 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T10:49:02,007 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=215 2025-01-08T10:49:02,007 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:49:02,007 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:49:02,007 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:49:02,007 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T10:49:02,009 DEBUG [HFileArchiver-21 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/e98b8450fec8e579da851a8d453681dd/cf/0d8d905ab1104abc8db75f006044a880 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportFileSystemStateWithSkipTmp/e98b8450fec8e579da851a8d453681dd/cf/0d8d905ab1104abc8db75f006044a880 2025-01-08T10:49:02,009 DEBUG [HFileArchiver-22 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/b2f5e0d7cfbb791b7590aef92c52e2ba/cf/bf223fa5ee0944ceae7efc0390fdbf63 to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportFileSystemStateWithSkipTmp/b2f5e0d7cfbb791b7590aef92c52e2ba/cf/bf223fa5ee0944ceae7efc0390fdbf63 2025-01-08T10:49:02,011 DEBUG [HFileArchiver-21 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/e98b8450fec8e579da851a8d453681dd/recovered.edits/9.seqid to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportFileSystemStateWithSkipTmp/e98b8450fec8e579da851a8d453681dd/recovered.edits/9.seqid 2025-01-08T10:49:02,012 DEBUG [HFileArchiver-22 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/b2f5e0d7cfbb791b7590aef92c52e2ba/recovered.edits/9.seqid to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportFileSystemStateWithSkipTmp/b2f5e0d7cfbb791b7590aef92c52e2ba/recovered.edits/9.seqid 2025-01-08T10:49:02,012 DEBUG [HFileArchiver-21 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/e98b8450fec8e579da851a8d453681dd 2025-01-08T10:49:02,012 DEBUG [HFileArchiver-22 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testtb-testExportFileSystemStateWithSkipTmp/b2f5e0d7cfbb791b7590aef92c52e2ba 2025-01-08T10:49:02,012 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(313): Archived testtb-testExportFileSystemStateWithSkipTmp regions 2025-01-08T10:49:02,012 DEBUG [PEWorker-5 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a 2025-01-08T10:49:02,013 DEBUG [PEWorker-5 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf] 2025-01-08T10:49:02,015 DEBUG [PEWorker-5 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/c4ca4238a0b923820dcc509a6f75849b20250108d6408b293fcd4556afd0a5cbb0d601fe_b2f5e0d7cfbb791b7590aef92c52e2ba to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/c4ca4238a0b923820dcc509a6f75849b20250108d6408b293fcd4556afd0a5cbb0d601fe_b2f5e0d7cfbb791b7590aef92c52e2ba 2025-01-08T10:49:02,016 DEBUG [PEWorker-5 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/d41d8cd98f00b204e9800998ecf8427e202501087a91e33d1d6a4a3abb5ec70e4a831205_e98b8450fec8e579da851a8d453681dd to hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/archive/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/d41d8cd98f00b204e9800998ecf8427e202501087a91e33d1d6a4a3abb5ec70e4a831205_e98b8450fec8e579da851a8d453681dd 2025-01-08T10:49:02,017 DEBUG [PEWorker-5 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a 2025-01-08T10:49:02,019 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=215, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T10:49:02,021 WARN [PEWorker-5 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testtb-testExportFileSystemStateWithSkipTmp from hbase:meta 2025-01-08T10:49:02,022 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testExportFileSystemStateWithSkipTmp' descriptor. 2025-01-08T10:49:02,023 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=215, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T10:49:02,023 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testExportFileSystemStateWithSkipTmp' from region states. 2025-01-08T10:49:02,023 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736333342023"}]},"ts":"9223372036854775807"} 2025-01-08T10:49:02,023 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736333342023"}]},"ts":"9223372036854775807"} 2025-01-08T10:49:02,025 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T10:49:02,025 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => e98b8450fec8e579da851a8d453681dd, NAME => 'testtb-testExportFileSystemStateWithSkipTmp,,1736333325633.e98b8450fec8e579da851a8d453681dd.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => b2f5e0d7cfbb791b7590aef92c52e2ba, NAME => 'testtb-testExportFileSystemStateWithSkipTmp,1,1736333325633.b2f5e0d7cfbb791b7590aef92c52e2ba.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T10:49:02,025 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testExportFileSystemStateWithSkipTmp' as deleted. 2025-01-08T10:49:02,025 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithSkipTmp","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736333342025"}]},"ts":"9223372036854775807"} 2025-01-08T10:49:02,027 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testExportFileSystemStateWithSkipTmp state from META 2025-01-08T10:49:02,029 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(133): Finished pid=215, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T10:49:02,029 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=215, state=SUCCESS; DeleteTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp in 32 msec 2025-01-08T10:49:02,108 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=215 2025-01-08T10:49:02,108 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testExportFileSystemStateWithSkipTmp, procId: 215 completed 2025-01-08T10:49:02,113 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.2 delete name: "emptySnaptb0-testExportFileSystemStateWithSkipTmp" 2025-01-08T10:49:02,115 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotManager(380): Deleting snapshot: emptySnaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T10:49:02,116 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.2 delete name: "snaptb0-testExportFileSystemStateWithSkipTmp" 2025-01-08T10:49:02,117 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T10:49:02,138 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: snapshot.TestMobSecureExportSnapshot#testExportFileSystemStateWithSkipTmp Thread=803 (was 799) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-50 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-52 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-22 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:36461 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_2013034284_1 at /127.0.0.1:51608 [Waiting for operation #3] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (175901922) connection to localhost/127.0.0.1:36461 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-832679797_22 at /127.0.0.1:38518 [Waiting for operation #3] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-21 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #17 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-832679797_22 at /127.0.0.1:48128 [Waiting for operation #5] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-53 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Thread-8007 java.base@17.0.11/java.io.FileInputStream.readBytes(Native Method) java.base@17.0.11/java.io.FileInputStream.read(FileInputStream.java:276) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:282) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) java.base@17.0.11/sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:281) java.base@17.0.11/sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:324) java.base@17.0.11/sun.nio.cs.StreamDecoder.read(StreamDecoder.java:189) java.base@17.0.11/java.io.InputStreamReader.read(InputStreamReader.java:177) java.base@17.0.11/java.io.BufferedReader.fill(BufferedReader.java:162) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:329) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:396) app//org.apache.hadoop.util.Shell$1.run(Shell.java:1025) Potentially hanging thread: hconnection-0x5fcb6b98-shared-pool-51 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #18 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-832679797_22 at /127.0.0.1:51650 [Waiting for operation #4] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: process reaper (pid 29643) java.base@17.0.11/java.lang.ProcessHandleImpl.waitForProcessExit0(Native Method) java.base@17.0.11/java.lang.ProcessHandleImpl$1.run(ProcessHandleImpl.java:150) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) - Thread LEAK? -, OpenFileDescriptor=804 (was 807), MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=375 (was 465), ProcessCount=18 (was 18), AvailableMemoryMB=562 (was 555) - AvailableMemoryMB LEAK? - 2025-01-08T10:49:02,138 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=803 is superior to 500 2025-01-08T10:49:02,138 INFO [Time-limited test {}] hbase.HBaseTestingUtility(2861): Stopping mini mapreduce cluster... 2025-01-08T10:49:02,145 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@2d6c849d{node,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/node} 2025-01-08T10:49:02,147 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@759bc9dd{HTTP/1.1, (http/1.1)}{localhost:0} 2025-01-08T10:49:02,148 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2025-01-08T10:49:02,148 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6868b947{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/static,STOPPED} 2025-01-08T10:49:02,148 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@5f820e1d{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop.log.dir/,STOPPED} 2025-01-08T10:49:02,162 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(360): Exit code from container container_1736333116683_0010_01_000001 is : 143 2025-01-08T10:49:02,173 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_1/usercache/jenkins/appcache/application_1736333116683_0010/container_1736333116683_0010_01_000001/launch_container.sh] 2025-01-08T10:49:02,173 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_1/usercache/jenkins/appcache/application_1736333116683_0010/container_1736333116683_0010_01_000001/container_tokens] 2025-01-08T10:49:02,173 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/MiniMRCluster_1816529756/yarn-203358626/MiniMRCluster_1816529756-localDir-nm-1_1/usercache/jenkins/appcache/application_1736333116683_0010/container_1736333116683_0010_01_000001/sysfs] 2025-01-08T10:49:03,010 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 5f4e73fdd950eac2bfc23f9cc1241d47, had cached 0 bytes from a total of 14859 2025-01-08T10:49:03,010 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region a8e8f2d8192bce47f7fedf056a7a63d9, had cached 0 bytes from a total of 5688 2025-01-08T10:49:05,715 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T10:49:07,390 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T10:49:07,595 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T10:49:13,097 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T10:49:19,160 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@f3ffbde{node,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/node} 2025-01-08T10:49:19,161 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@61329307{HTTP/1.1, (http/1.1)}{localhost:0} 2025-01-08T10:49:19,161 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2025-01-08T10:49:19,161 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@47e31f8f{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/static,STOPPED} 2025-01-08T10:49:19,161 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@10e17038{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop.log.dir/,STOPPED} 2025-01-08T10:49:35,715 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T10:49:36,168 ERROR [Thread[Thread-418,5,FailOnTimeoutGroup] {}] delegation.AbstractDelegationTokenSecretManager$ExpiredTokenRemover(852): ExpiredTokenRemover received java.lang.InterruptedException: sleep interrupted 2025-01-08T10:49:36,169 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@7c1fb4a6{cluster,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/cluster} 2025-01-08T10:49:36,169 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@3aa3cd9c{HTTP/1.1, (http/1.1)}{localhost:0} 2025-01-08T10:49:36,169 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2025-01-08T10:49:36,169 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@fc1b748{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/static,STOPPED} 2025-01-08T10:49:36,169 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@5d5a85fc{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop.log.dir/,STOPPED} 2025-01-08T10:49:36,173 WARN [ApplicationMaster Launcher {}] amlauncher.ApplicationMasterLauncher$LauncherThread(122): org.apache.hadoop.yarn.server.resourcemanager.amlauncher.ApplicationMasterLauncher$LauncherThread interrupted. Returning. 2025-01-08T10:49:36,178 ERROR [SchedulerEventDispatcher:Event Processor {}] event.EventDispatcher$EventProcessor(72): Returning, interrupted : java.lang.InterruptedException 2025-01-08T10:49:36,178 ERROR [ResourceManager Event Processor Monitor {}] resourcemanager.ResourceManager$SchedulerEventDispatcher$EventProcessorMonitor(1193): Returning, interrupted : java.lang.InterruptedException: sleep interrupted 2025-01-08T10:49:36,181 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741830_1006 (size=952486) 2025-01-08T10:49:36,181 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741830_1006 (size=952486) 2025-01-08T10:49:36,181 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741830_1006 (size=952486) 2025-01-08T10:49:36,182 ERROR [Thread[Thread-442,5,FailOnTimeoutGroup] {}] delegation.AbstractDelegationTokenSecretManager$ExpiredTokenRemover(852): ExpiredTokenRemover received java.lang.InterruptedException: sleep interrupted 2025-01-08T10:49:36,186 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@7b7fab8{jobhistory,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/jobhistory} 2025-01-08T10:49:36,187 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@7c1e8a21{HTTP/1.1, (http/1.1)}{localhost:0} 2025-01-08T10:49:36,187 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2025-01-08T10:49:36,187 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3a64ffa4{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/static,STOPPED} 2025-01-08T10:49:36,187 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@431bcc59{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop.log.dir/,STOPPED} 2025-01-08T10:49:36,189 ERROR [Thread[Thread-400,5,FailOnTimeoutGroup] {}] delegation.AbstractDelegationTokenSecretManager$ExpiredTokenRemover(852): ExpiredTokenRemover received java.lang.InterruptedException: sleep interrupted 2025-01-08T10:49:36,189 INFO [Time-limited test {}] hbase.HBaseTestingUtility(2864): Mini mapreduce cluster stopped 2025-01-08T10:49:36,189 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1340): Shutting down minicluster 2025-01-08T10:49:36,189 INFO [Time-limited test {}] client.ConnectionImplementation(2127): Closing master protocol: MasterService 2025-01-08T10:49:36,189 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x067db5ed to 127.0.0.1:52367 2025-01-08T10:49:36,189 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:49:36,189 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2025-01-08T10:49:36,189 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=354350775, stopped=false 2025-01-08T10:49:36,190 DEBUG [Time-limited test {}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:49:36,190 DEBUG [Time-limited test {}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.SecureTestUtil$MasterSyncObserver 2025-01-08T10:49:36,190 INFO [Time-limited test {}] master.ServerManager(987): Cluster shutdown requested of master=67f42ee69b4e,34551,1736333106957 2025-01-08T10:49:36,191 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2025-01-08T10:49:36,191 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2025-01-08T10:49:36,191 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:49:36,191 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2025-01-08T10:49:36,191 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2025-01-08T10:49:36,192 INFO [Time-limited test {}] procedure2.ProcedureExecutor(700): Stopping 2025-01-08T10:49:36,192 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:49:36,192 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:49:36,192 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:49:36,192 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2025-01-08T10:49:36,192 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2025-01-08T10:49:36,192 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:49:36,192 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2025-01-08T10:49:36,193 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2025-01-08T10:49:36,193 INFO [Time-limited test {}] regionserver.HRegionServer(2561): ***** STOPPING region server '67f42ee69b4e,45003,1736333108039' ***** 2025-01-08T10:49:36,193 DEBUG [Time-limited test {}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:49:36,193 INFO [Time-limited test {}] regionserver.HRegionServer(2575): STOPPED: Shutdown requested 2025-01-08T10:49:36,193 INFO [Time-limited test {}] regionserver.HRegionServer(2561): ***** STOPPING region server '67f42ee69b4e,36283,1736333108166' ***** 2025-01-08T10:49:36,193 DEBUG [Time-limited test {}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:49:36,193 INFO [Time-limited test {}] regionserver.HRegionServer(2575): STOPPED: Shutdown requested 2025-01-08T10:49:36,193 INFO [Time-limited test {}] regionserver.HRegionServer(2561): ***** STOPPING region server '67f42ee69b4e,38323,1736333108259' ***** 2025-01-08T10:49:36,193 DEBUG [Time-limited test {}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:49:36,193 INFO [RS:0;67f42ee69b4e:45003 {}] regionserver.HeapMemoryManager(220): Stopping 2025-01-08T10:49:36,193 INFO [Time-limited test {}] regionserver.HRegionServer(2575): STOPPED: Shutdown requested 2025-01-08T10:49:36,193 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(373): MemStoreFlusher.0 exiting 2025-01-08T10:49:36,193 INFO [RS:1;67f42ee69b4e:36283 {}] regionserver.HeapMemoryManager(220): Stopping 2025-01-08T10:49:36,193 INFO [RS:0;67f42ee69b4e:45003 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2025-01-08T10:49:36,193 INFO [RS:1;67f42ee69b4e:36283 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2025-01-08T10:49:36,193 INFO [RS:0;67f42ee69b4e:45003 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2025-01-08T10:49:36,193 INFO [RS:1;67f42ee69b4e:36283 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2025-01-08T10:49:36,193 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(373): MemStoreFlusher.0 exiting 2025-01-08T10:49:36,194 INFO [RS:2;67f42ee69b4e:38323 {}] regionserver.HeapMemoryManager(220): Stopping 2025-01-08T10:49:36,194 INFO [RS:2;67f42ee69b4e:38323 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2025-01-08T10:49:36,194 INFO [RS:2;67f42ee69b4e:38323 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2025-01-08T10:49:36,194 INFO [RS:1;67f42ee69b4e:36283 {}] regionserver.HRegionServer(3579): Received CLOSE for a8e8f2d8192bce47f7fedf056a7a63d9 2025-01-08T10:49:36,194 INFO [RS:0;67f42ee69b4e:45003 {}] regionserver.HRegionServer(3579): Received CLOSE for 5f4e73fdd950eac2bfc23f9cc1241d47 2025-01-08T10:49:36,194 INFO [RS:2;67f42ee69b4e:38323 {}] regionserver.HRegionServer(3579): Received CLOSE for 873898d3d33a2f6df797594f02de3737 2025-01-08T10:49:36,194 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(373): MemStoreFlusher.0 exiting 2025-01-08T10:49:36,194 INFO [RS:0;67f42ee69b4e:45003 {}] regionserver.HRegionServer(1224): stopping server 67f42ee69b4e,45003,1736333108039 2025-01-08T10:49:36,194 INFO [RS:2;67f42ee69b4e:38323 {}] regionserver.HRegionServer(1224): stopping server 67f42ee69b4e,38323,1736333108259 2025-01-08T10:49:36,194 INFO [RS:1;67f42ee69b4e:36283 {}] regionserver.HRegionServer(3579): Received CLOSE for 0c62df083bd10295a5e532128fcb0af8 2025-01-08T10:49:36,194 DEBUG [RS:0;67f42ee69b4e:45003 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:49:36,194 DEBUG [RS:2;67f42ee69b4e:38323 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:49:36,194 INFO [RS:1;67f42ee69b4e:36283 {}] regionserver.HRegionServer(1224): stopping server 67f42ee69b4e,36283,1736333108166 2025-01-08T10:49:36,194 DEBUG [RS:1;67f42ee69b4e:36283 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:49:36,194 INFO [RS:0;67f42ee69b4e:45003 {}] regionserver.HRegionServer(1599): Waiting on 1 regions to close 2025-01-08T10:49:36,194 INFO [RS:2;67f42ee69b4e:38323 {}] regionserver.HRegionServer(1599): Waiting on 1 regions to close 2025-01-08T10:49:36,194 DEBUG [RS:0;67f42ee69b4e:45003 {}] regionserver.HRegionServer(1603): Online Regions={5f4e73fdd950eac2bfc23f9cc1241d47=testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47.} 2025-01-08T10:49:36,194 DEBUG [RS:2;67f42ee69b4e:38323 {}] regionserver.HRegionServer(1603): Online Regions={873898d3d33a2f6df797594f02de3737=hbase:namespace,,1736333110618.873898d3d33a2f6df797594f02de3737.} 2025-01-08T10:49:36,194 INFO [RS:1;67f42ee69b4e:36283 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2025-01-08T10:49:36,194 INFO [RS:1;67f42ee69b4e:36283 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2025-01-08T10:49:36,194 INFO [RS:1;67f42ee69b4e:36283 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2025-01-08T10:49:36,194 INFO [RS:1;67f42ee69b4e:36283 {}] regionserver.HRegionServer(3579): Received CLOSE for 1588230740 2025-01-08T10:49:36,194 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 873898d3d33a2f6df797594f02de3737, disabling compactions & flushes 2025-01-08T10:49:36,195 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region hbase:namespace,,1736333110618.873898d3d33a2f6df797594f02de3737. 2025-01-08T10:49:36,195 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1736333110618.873898d3d33a2f6df797594f02de3737. 2025-01-08T10:49:36,195 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1736333110618.873898d3d33a2f6df797594f02de3737. after waiting 0 ms 2025-01-08T10:49:36,195 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1736333110618.873898d3d33a2f6df797594f02de3737. 2025-01-08T10:49:36,195 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(2837): Flushing 873898d3d33a2f6df797594f02de3737 1/1 column families, dataSize=78 B heapSize=488 B 2025-01-08T10:49:36,197 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 5f4e73fdd950eac2bfc23f9cc1241d47, disabling compactions & flushes 2025-01-08T10:49:36,197 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47. 2025-01-08T10:49:36,197 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47. 2025-01-08T10:49:36,197 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47. after waiting 0 ms 2025-01-08T10:49:36,197 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47. 2025-01-08T10:49:36,198 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing a8e8f2d8192bce47f7fedf056a7a63d9, disabling compactions & flushes 2025-01-08T10:49:36,198 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9. 2025-01-08T10:49:36,198 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9. 2025-01-08T10:49:36,198 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9. after waiting 0 ms 2025-01-08T10:49:36,198 DEBUG [RS:0;67f42ee69b4e:45003 {}] regionserver.HRegionServer(1629): Waiting on 5f4e73fdd950eac2bfc23f9cc1241d47 2025-01-08T10:49:36,198 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9. 2025-01-08T10:49:36,198 DEBUG [RS:2;67f42ee69b4e:38323 {}] regionserver.HRegionServer(1629): Waiting on 873898d3d33a2f6df797594f02de3737 2025-01-08T10:49:36,198 INFO [RS:1;67f42ee69b4e:36283 {}] regionserver.HRegionServer(1599): Waiting on 3 regions to close 2025-01-08T10:49:36,198 DEBUG [RS:1;67f42ee69b4e:36283 {}] regionserver.HRegionServer(1603): Online Regions={a8e8f2d8192bce47f7fedf056a7a63d9=testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9., 1588230740=hbase:meta,,1.1588230740, 0c62df083bd10295a5e532128fcb0af8=hbase:acl,,1736333111799.0c62df083bd10295a5e532128fcb0af8.} 2025-01-08T10:49:36,198 DEBUG [RS:1;67f42ee69b4e:36283 {}] regionserver.HRegionServer(1629): Waiting on 0c62df083bd10295a5e532128fcb0af8, 1588230740, a8e8f2d8192bce47f7fedf056a7a63d9 2025-01-08T10:49:36,198 DEBUG [RS_CLOSE_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2025-01-08T10:49:36,198 INFO [RS_CLOSE_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2025-01-08T10:49:36,198 DEBUG [RS_CLOSE_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2025-01-08T10:49:36,198 DEBUG [RS_CLOSE_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2025-01-08T10:49:36,198 DEBUG [RS_CLOSE_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2025-01-08T10:49:36,199 INFO [RS_CLOSE_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2837): Flushing 1588230740 3/3 column families, dataSize=68.66 KB heapSize=109 KB 2025-01-08T10:49:36,207 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportExpiredSnapshot/5f4e73fdd950eac2bfc23f9cc1241d47/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2025-01-08T10:49:36,208 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:49:36,208 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47. 2025-01-08T10:49:36,208 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 5f4e73fdd950eac2bfc23f9cc1241d47: 2025-01-08T10:49:36,208 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47. 2025-01-08T10:49:36,208 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/default/testExportExpiredSnapshot/a8e8f2d8192bce47f7fedf056a7a63d9/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2025-01-08T10:49:36,209 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:49:36,209 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9. 2025-01-08T10:49:36,209 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for a8e8f2d8192bce47f7fedf056a7a63d9: 2025-01-08T10:49:36,209 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed testExportExpiredSnapshot,,1736333252601.a8e8f2d8192bce47f7fedf056a7a63d9. 2025-01-08T10:49:36,209 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 0c62df083bd10295a5e532128fcb0af8, disabling compactions & flushes 2025-01-08T10:49:36,209 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region hbase:acl,,1736333111799.0c62df083bd10295a5e532128fcb0af8. 2025-01-08T10:49:36,209 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:acl,,1736333111799.0c62df083bd10295a5e532128fcb0af8. 2025-01-08T10:49:36,209 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on hbase:acl,,1736333111799.0c62df083bd10295a5e532128fcb0af8. after waiting 0 ms 2025-01-08T10:49:36,209 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region hbase:acl,,1736333111799.0c62df083bd10295a5e532128fcb0af8. 2025-01-08T10:49:36,209 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(2837): Flushing 0c62df083bd10295a5e532128fcb0af8 1/1 column families, dataSize=1.38 KB heapSize=3.33 KB 2025-01-08T10:49:36,213 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/namespace/873898d3d33a2f6df797594f02de3737/.tmp/info/ac662fb9f26d475e85582912a5c859a3 is 45, key is default/info:d/1736333111714/Put/seqid=0 2025-01-08T10:49:36,220 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742431_1607 (size=5037) 2025-01-08T10:49:36,220 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742431_1607 (size=5037) 2025-01-08T10:49:36,220 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742431_1607 (size=5037) 2025-01-08T10:49:36,221 INFO [regionserver/67f42ee69b4e:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2025-01-08T10:49:36,221 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=78 B at sequenceid=6 (bloomFilter=true), to=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/namespace/873898d3d33a2f6df797594f02de3737/.tmp/info/ac662fb9f26d475e85582912a5c859a3 2025-01-08T10:49:36,225 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/namespace/873898d3d33a2f6df797594f02de3737/.tmp/info/ac662fb9f26d475e85582912a5c859a3 as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/namespace/873898d3d33a2f6df797594f02de3737/info/ac662fb9f26d475e85582912a5c859a3 2025-01-08T10:49:36,229 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/namespace/873898d3d33a2f6df797594f02de3737/info/ac662fb9f26d475e85582912a5c859a3, entries=2, sequenceid=6, filesize=4.9 K 2025-01-08T10:49:36,229 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/acl/0c62df083bd10295a5e532128fcb0af8/.tmp/l/973eae44f3e04c1a8039809ded1a518c is 74, key is testtb-testExportFileSystemStateWithMergeRegion-1/l:/1736333249911/DeleteFamily/seqid=0 2025-01-08T10:49:36,230 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(3040): Finished flush of dataSize ~78 B/78, heapSize ~472 B/472, currentSize=0 B/0 for 873898d3d33a2f6df797594f02de3737 in 34ms, sequenceid=6, compaction requested=false 2025-01-08T10:49:36,233 DEBUG [RS_CLOSE_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/meta/1588230740/.tmp/info/cce7c755042f4446a38a9a54d26273a2 is 173, key is testExportExpiredSnapshot,1,1736333252601.5f4e73fdd950eac2bfc23f9cc1241d47./info:regioninfo/1736333253027/Put/seqid=0 2025-01-08T10:49:36,235 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/namespace/873898d3d33a2f6df797594f02de3737/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T10:49:36,235 INFO [regionserver/67f42ee69b4e:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2025-01-08T10:49:36,236 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:49:36,236 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed hbase:namespace,,1736333110618.873898d3d33a2f6df797594f02de3737. 2025-01-08T10:49:36,236 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 873898d3d33a2f6df797594f02de3737: 2025-01-08T10:49:36,236 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed hbase:namespace,,1736333110618.873898d3d33a2f6df797594f02de3737. 2025-01-08T10:49:36,236 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742432_1608 (size=5695) 2025-01-08T10:49:36,236 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742432_1608 (size=5695) 2025-01-08T10:49:36,237 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742432_1608 (size=5695) 2025-01-08T10:49:36,237 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.38 KB at sequenceid=27 (bloomFilter=false), to=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/acl/0c62df083bd10295a5e532128fcb0af8/.tmp/l/973eae44f3e04c1a8039809ded1a518c 2025-01-08T10:49:36,238 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742433_1609 (size=15630) 2025-01-08T10:49:36,239 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742433_1609 (size=15630) 2025-01-08T10:49:36,239 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742433_1609 (size=15630) 2025-01-08T10:49:36,240 INFO [RS_CLOSE_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=60.26 KB at sequenceid=202 (bloomFilter=true), to=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/meta/1588230740/.tmp/info/cce7c755042f4446a38a9a54d26273a2 2025-01-08T10:49:36,241 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.StoreFileReader(539): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for 973eae44f3e04c1a8039809ded1a518c 2025-01-08T10:49:36,242 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/acl/0c62df083bd10295a5e532128fcb0af8/.tmp/l/973eae44f3e04c1a8039809ded1a518c as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/acl/0c62df083bd10295a5e532128fcb0af8/l/973eae44f3e04c1a8039809ded1a518c 2025-01-08T10:49:36,242 INFO [regionserver/67f42ee69b4e:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2025-01-08T10:49:36,245 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.StoreFileReader(539): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for 973eae44f3e04c1a8039809ded1a518c 2025-01-08T10:49:36,245 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/acl/0c62df083bd10295a5e532128fcb0af8/l/973eae44f3e04c1a8039809ded1a518c, entries=12, sequenceid=27, filesize=5.6 K 2025-01-08T10:49:36,246 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(3040): Finished flush of dataSize ~1.38 KB/1412, heapSize ~3.31 KB/3392, currentSize=0 B/0 for 0c62df083bd10295a5e532128fcb0af8 in 37ms, sequenceid=27, compaction requested=false 2025-01-08T10:49:36,249 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/acl/0c62df083bd10295a5e532128fcb0af8/recovered.edits/30.seqid, newMaxSeqId=30, maxSeqId=1 2025-01-08T10:49:36,249 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:49:36,249 INFO [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed hbase:acl,,1736333111799.0c62df083bd10295a5e532128fcb0af8. 2025-01-08T10:49:36,249 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 0c62df083bd10295a5e532128fcb0af8: 2025-01-08T10:49:36,249 DEBUG [RS_CLOSE_REGION-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed hbase:acl,,1736333111799.0c62df083bd10295a5e532128fcb0af8. 2025-01-08T10:49:36,259 DEBUG [RS_CLOSE_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/meta/1588230740/.tmp/rep_barrier/80a00899e96342dba63225e108307d8a is 133, key is testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227962.96d4bb39d5de2c0741f205ac02281c75./rep_barrier:/1736333249936/DeleteFamily/seqid=0 2025-01-08T10:49:36,268 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742434_1610 (size=8007) 2025-01-08T10:49:36,268 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742434_1610 (size=8007) 2025-01-08T10:49:36,268 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742434_1610 (size=8007) 2025-01-08T10:49:36,268 INFO [RS_CLOSE_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=2.34 KB at sequenceid=202 (bloomFilter=true), to=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/meta/1588230740/.tmp/rep_barrier/80a00899e96342dba63225e108307d8a 2025-01-08T10:49:36,287 DEBUG [RS_CLOSE_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/meta/1588230740/.tmp/table/a42860362ab240bf9d7b769c37f06a51 is 127, key is testtb-testExportFileSystemStateWithMergeRegion-1,,1736333227962.96d4bb39d5de2c0741f205ac02281c75./table:/1736333249936/DeleteFamily/seqid=0 2025-01-08T10:49:36,291 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073742435_1611 (size=8861) 2025-01-08T10:49:36,291 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073742435_1611 (size=8861) 2025-01-08T10:49:36,292 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073742435_1611 (size=8861) 2025-01-08T10:49:36,292 INFO [RS_CLOSE_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=6.06 KB at sequenceid=202 (bloomFilter=true), to=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/meta/1588230740/.tmp/table/a42860362ab240bf9d7b769c37f06a51 2025-01-08T10:49:36,296 DEBUG [RS_CLOSE_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/meta/1588230740/.tmp/info/cce7c755042f4446a38a9a54d26273a2 as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/meta/1588230740/info/cce7c755042f4446a38a9a54d26273a2 2025-01-08T10:49:36,300 INFO [RS_CLOSE_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/meta/1588230740/info/cce7c755042f4446a38a9a54d26273a2, entries=84, sequenceid=202, filesize=15.3 K 2025-01-08T10:49:36,300 DEBUG [RS_CLOSE_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/meta/1588230740/.tmp/rep_barrier/80a00899e96342dba63225e108307d8a as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/meta/1588230740/rep_barrier/80a00899e96342dba63225e108307d8a 2025-01-08T10:49:36,304 INFO [RS_CLOSE_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/meta/1588230740/rep_barrier/80a00899e96342dba63225e108307d8a, entries=21, sequenceid=202, filesize=7.8 K 2025-01-08T10:49:36,304 DEBUG [RS_CLOSE_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/meta/1588230740/.tmp/table/a42860362ab240bf9d7b769c37f06a51 as hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/meta/1588230740/table/a42860362ab240bf9d7b769c37f06a51 2025-01-08T10:49:36,308 INFO [RS_CLOSE_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/meta/1588230740/table/a42860362ab240bf9d7b769c37f06a51, entries=38, sequenceid=202, filesize=8.7 K 2025-01-08T10:49:36,309 INFO [RS_CLOSE_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3040): Finished flush of dataSize ~68.66 KB/70312, heapSize ~108.95 KB/111568, currentSize=0 B/0 for 1588230740 in 111ms, sequenceid=202, compaction requested=false 2025-01-08T10:49:36,312 DEBUG [RS_CLOSE_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/data/hbase/meta/1588230740/recovered.edits/205.seqid, newMaxSeqId=205, maxSeqId=1 2025-01-08T10:49:36,313 DEBUG [RS_CLOSE_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:49:36,313 DEBUG [RS_CLOSE_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2025-01-08T10:49:36,313 INFO [RS_CLOSE_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2025-01-08T10:49:36,313 DEBUG [RS_CLOSE_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1635): Region close journal for 1588230740: 2025-01-08T10:49:36,313 DEBUG [RS_CLOSE_META-regionserver/67f42ee69b4e:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2025-01-08T10:49:36,398 INFO [RS:0;67f42ee69b4e:45003 {}] regionserver.HRegionServer(1250): stopping server 67f42ee69b4e,45003,1736333108039; all regions closed. 2025-01-08T10:49:36,398 INFO [RS:2;67f42ee69b4e:38323 {}] regionserver.HRegionServer(1250): stopping server 67f42ee69b4e,38323,1736333108259; all regions closed. 2025-01-08T10:49:36,398 INFO [RS:1;67f42ee69b4e:36283 {}] regionserver.HRegionServer(1250): stopping server 67f42ee69b4e,36283,1736333108166; all regions closed. 2025-01-08T10:49:36,402 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741835_1011 (size=11814) 2025-01-08T10:49:36,402 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741836_1012 (size=80694) 2025-01-08T10:49:36,402 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741835_1011 (size=11814) 2025-01-08T10:49:36,403 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741835_1011 (size=11814) 2025-01-08T10:49:36,403 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741834_1010 (size=10789) 2025-01-08T10:49:36,403 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741834_1010 (size=10789) 2025-01-08T10:49:36,403 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741834_1010 (size=10789) 2025-01-08T10:49:36,403 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741836_1012 (size=80694) 2025-01-08T10:49:36,403 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741836_1012 (size=80694) 2025-01-08T10:49:36,405 DEBUG [RS:1;67f42ee69b4e:36283 {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/oldWALs 2025-01-08T10:49:36,405 INFO [RS:1;67f42ee69b4e:36283 {}] wal.AbstractFSWAL(1074): Closed WAL: AsyncFSWAL 67f42ee69b4e%2C36283%2C1736333108166.meta:.meta(num 1736333110392) 2025-01-08T10:49:36,405 DEBUG [RS:0;67f42ee69b4e:45003 {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/oldWALs 2025-01-08T10:49:36,405 INFO [RS:0;67f42ee69b4e:45003 {}] wal.AbstractFSWAL(1074): Closed WAL: AsyncFSWAL 67f42ee69b4e%2C45003%2C1736333108039:(num 1736333109986) 2025-01-08T10:49:36,405 DEBUG [RS:0;67f42ee69b4e:45003 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:49:36,405 INFO [RS:0;67f42ee69b4e:45003 {}] regionserver.LeaseManager(133): Closed leases 2025-01-08T10:49:36,406 INFO [RS:0;67f42ee69b4e:45003 {}] hbase.ChoreService(370): Chore service for: regionserver/67f42ee69b4e:0 had [ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS] on shutdown 2025-01-08T10:49:36,406 INFO [RS:0;67f42ee69b4e:45003 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2025-01-08T10:49:36,406 INFO [RS:0;67f42ee69b4e:45003 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2025-01-08T10:49:36,406 INFO [RS:0;67f42ee69b4e:45003 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2025-01-08T10:49:36,406 DEBUG [RS:2;67f42ee69b4e:38323 {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/oldWALs 2025-01-08T10:49:36,406 INFO [RS:2;67f42ee69b4e:38323 {}] wal.AbstractFSWAL(1074): Closed WAL: AsyncFSWAL 67f42ee69b4e%2C38323%2C1736333108259:(num 1736333109984) 2025-01-08T10:49:36,406 DEBUG [RS:2;67f42ee69b4e:38323 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:49:36,406 INFO [RS:2;67f42ee69b4e:38323 {}] regionserver.LeaseManager(133): Closed leases 2025-01-08T10:49:36,406 INFO [RS:2;67f42ee69b4e:38323 {}] hbase.ChoreService(370): Chore service for: regionserver/67f42ee69b4e:0 had [ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS] on shutdown 2025-01-08T10:49:36,406 INFO [regionserver/67f42ee69b4e:0.logRoller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2025-01-08T10:49:36,407 INFO [RS:2;67f42ee69b4e:38323 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2025-01-08T10:49:36,407 INFO [RS:0;67f42ee69b4e:45003 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:45003 2025-01-08T10:49:36,407 INFO [RS:2;67f42ee69b4e:38323 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2025-01-08T10:49:36,407 INFO [RS:2;67f42ee69b4e:38323 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2025-01-08T10:49:36,408 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45285 is added to blk_1073741833_1009 (size=15671) 2025-01-08T10:49:36,408 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44473 is added to blk_1073741833_1009 (size=15671) 2025-01-08T10:49:36,408 INFO [RS:2;67f42ee69b4e:38323 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:38323 2025-01-08T10:49:36,409 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43487 is added to blk_1073741833_1009 (size=15671) 2025-01-08T10:49:36,410 INFO [regionserver/67f42ee69b4e:0.logRoller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2025-01-08T10:49:36,411 DEBUG [RS:1;67f42ee69b4e:36283 {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/oldWALs 2025-01-08T10:49:36,411 INFO [RS:1;67f42ee69b4e:36283 {}] wal.AbstractFSWAL(1074): Closed WAL: AsyncFSWAL 67f42ee69b4e%2C36283%2C1736333108166:(num 1736333109982) 2025-01-08T10:49:36,411 DEBUG [RS:1;67f42ee69b4e:36283 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:49:36,411 INFO [RS:1;67f42ee69b4e:36283 {}] regionserver.LeaseManager(133): Closed leases 2025-01-08T10:49:36,411 INFO [RS:1;67f42ee69b4e:36283 {}] hbase.ChoreService(370): Chore service for: regionserver/67f42ee69b4e:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS] on shutdown 2025-01-08T10:49:36,412 INFO [regionserver/67f42ee69b4e:0.logRoller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2025-01-08T10:49:36,412 INFO [RS:1;67f42ee69b4e:36283 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:36283 2025-01-08T10:49:36,412 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2025-01-08T10:49:36,412 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/67f42ee69b4e,38323,1736333108259 2025-01-08T10:49:36,413 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/67f42ee69b4e,45003,1736333108039 2025-01-08T10:49:36,415 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/67f42ee69b4e,36283,1736333108166 2025-01-08T10:49:36,415 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [67f42ee69b4e,36283,1736333108166] 2025-01-08T10:49:36,415 DEBUG [RegionServerTracker-0 {}] master.DeadServer(103): Processing 67f42ee69b4e,36283,1736333108166; numProcessing=1 2025-01-08T10:49:36,416 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/draining/67f42ee69b4e,36283,1736333108166 already deleted, retry=false 2025-01-08T10:49:36,416 INFO [RegionServerTracker-0 {}] master.ServerManager(652): Cluster shutdown set; 67f42ee69b4e,36283,1736333108166 expired; onlineServers=2 2025-01-08T10:49:36,416 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [67f42ee69b4e,45003,1736333108039] 2025-01-08T10:49:36,416 DEBUG [RegionServerTracker-0 {}] master.DeadServer(103): Processing 67f42ee69b4e,45003,1736333108039; numProcessing=2 2025-01-08T10:49:36,418 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/draining/67f42ee69b4e,45003,1736333108039 already deleted, retry=false 2025-01-08T10:49:36,418 INFO [RegionServerTracker-0 {}] master.ServerManager(652): Cluster shutdown set; 67f42ee69b4e,45003,1736333108039 expired; onlineServers=1 2025-01-08T10:49:36,418 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [67f42ee69b4e,38323,1736333108259] 2025-01-08T10:49:36,418 DEBUG [RegionServerTracker-0 {}] master.DeadServer(103): Processing 67f42ee69b4e,38323,1736333108259; numProcessing=3 2025-01-08T10:49:36,419 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/draining/67f42ee69b4e,38323,1736333108259 already deleted, retry=false 2025-01-08T10:49:36,419 INFO [RegionServerTracker-0 {}] master.ServerManager(652): Cluster shutdown set; 67f42ee69b4e,38323,1736333108259 expired; onlineServers=0 2025-01-08T10:49:36,419 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2561): ***** STOPPING region server '67f42ee69b4e,34551,1736333106957' ***** 2025-01-08T10:49:36,419 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2575): STOPPED: Cluster shutdown set; onlineServer=0 2025-01-08T10:49:36,419 DEBUG [M:0;67f42ee69b4e:34551 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@413a950f, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=67f42ee69b4e/172.17.0.2:0 2025-01-08T10:49:36,419 INFO [M:0;67f42ee69b4e:34551 {}] regionserver.HRegionServer(1224): stopping server 67f42ee69b4e,34551,1736333106957 2025-01-08T10:49:36,419 INFO [M:0;67f42ee69b4e:34551 {}] regionserver.HRegionServer(1250): stopping server 67f42ee69b4e,34551,1736333106957; all regions closed. 2025-01-08T10:49:36,419 DEBUG [M:0;67f42ee69b4e:34551 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T10:49:36,419 DEBUG [M:0;67f42ee69b4e:34551 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2025-01-08T10:49:36,419 DEBUG [M:0;67f42ee69b4e:34551 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2025-01-08T10:49:36,419 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2025-01-08T10:49:36,420 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster-HFileCleaner.large.0-1736333109662 {}] cleaner.HFileCleaner(306): Exit Thread[master/67f42ee69b4e:0:becomeActiveMaster-HFileCleaner.large.0-1736333109662,5,FailOnTimeoutGroup] 2025-01-08T10:49:36,420 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster-HFileCleaner.small.0-1736333109667 {}] cleaner.HFileCleaner(306): Exit Thread[master/67f42ee69b4e:0:becomeActiveMaster-HFileCleaner.small.0-1736333109667,5,FailOnTimeoutGroup] 2025-01-08T10:49:36,420 INFO [M:0;67f42ee69b4e:34551 {}] hbase.ChoreService(370): Chore service for: master/67f42ee69b4e:0 had [] on shutdown 2025-01-08T10:49:36,420 DEBUG [M:0;67f42ee69b4e:34551 {}] master.HMaster(1744): Stopping service threads 2025-01-08T10:49:36,420 INFO [M:0;67f42ee69b4e:34551 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2025-01-08T10:49:36,420 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2025-01-08T10:49:36,420 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T10:49:36,421 INFO [M:0;67f42ee69b4e:34551 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2025-01-08T10:49:36,421 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2025-01-08T10:49:36,421 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2025-01-08T10:49:36,421 DEBUG [M:0;67f42ee69b4e:34551 {}] zookeeper.ZKUtil(347): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2025-01-08T10:49:36,421 WARN [M:0;67f42ee69b4e:34551 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2025-01-08T10:49:36,421 INFO [M:0;67f42ee69b4e:34551 {}] assignment.AssignmentManager(391): Stopping assignment manager 2025-01-08T10:49:36,422 INFO [M:0;67f42ee69b4e:34551 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2025-01-08T10:49:36,422 DEBUG [M:0;67f42ee69b4e:34551 {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2025-01-08T10:49:36,437 INFO [M:0;67f42ee69b4e:34551 {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2025-01-08T10:49:36,437 DEBUG [M:0;67f42ee69b4e:34551 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2025-01-08T10:49:36,437 DEBUG [M:0;67f42ee69b4e:34551 {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2025-01-08T10:49:36,437 DEBUG [M:0;67f42ee69b4e:34551 {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2025-01-08T10:49:36,437 INFO [M:0;67f42ee69b4e:34551 {}] regionserver.HRegion(2837): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=810.90 KB heapSize=972.24 KB 2025-01-08T10:49:36,437 ERROR [AsyncFSWAL-0-hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/MasterData-prefix:67f42ee69b4e,34551,1736333106957 {}] server.NIOServerCnxnFactory(85): Thread Thread[AsyncFSWAL-0-hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/MasterData-prefix:67f42ee69b4e,34551,1736333106957,5,FailOnTimeoutGroup] died java.lang.NullPointerException: Cannot invoke "org.apache.hbase.thirdparty.io.netty.buffer.ByteBuf.readableBytes()" because "this.buf" is null at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutput.buffered(FanOutOneBlockAsyncDFSOutput.java:419) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.wal.AsyncProtobufLogWriter.append(AsyncProtobufLogWriter.java:132) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.doAppend(AsyncFSWAL.java:830) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.doAppend(AsyncFSWAL.java:128) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.appendEntry(AbstractFSWAL.java:1148) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.appendAndSync(AsyncFSWAL.java:500) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.consume(AsyncFSWAL.java:603) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2025-01-08T10:49:36,514 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2025-01-08T10:49:36,514 INFO [RS:2;67f42ee69b4e:38323 {}] regionserver.HRegionServer(1307): Exiting; stopping=67f42ee69b4e,38323,1736333108259; zookeeper connection closed. 2025-01-08T10:49:36,514 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38323-0x1000c1ee94b0003, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2025-01-08T10:49:36,515 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@4c4ebf25 {}] hbase.MiniHBaseCluster$SingleFileSystemShutdownThread(216): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@4c4ebf25 2025-01-08T10:49:36,515 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2025-01-08T10:49:36,515 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45003-0x1000c1ee94b0001, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2025-01-08T10:49:36,515 INFO [RS:0;67f42ee69b4e:45003 {}] regionserver.HRegionServer(1307): Exiting; stopping=67f42ee69b4e,45003,1736333108039; zookeeper connection closed. 2025-01-08T10:49:36,515 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@29f28ef5 {}] hbase.MiniHBaseCluster$SingleFileSystemShutdownThread(216): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@29f28ef5 2025-01-08T10:49:36,517 INFO [RS:1;67f42ee69b4e:36283 {}] regionserver.HRegionServer(1307): Exiting; stopping=67f42ee69b4e,36283,1736333108166; zookeeper connection closed. 2025-01-08T10:49:36,517 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2025-01-08T10:49:36,517 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36283-0x1000c1ee94b0002, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2025-01-08T10:49:36,517 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@55f22325 {}] hbase.MiniHBaseCluster$SingleFileSystemShutdownThread(216): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@55f22325 2025-01-08T10:49:36,517 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2025-01-08T10:49:37,595 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=Coprocessor.RegionServer.CP_org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:49:37,596 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_meta 2025-01-08T10:49:37,596 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_namespace 2025-01-08T10:49:37,596 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_acl 2025-01-08T10:49:37,596 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: Master,sub=Coprocessor.Master.CP_org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:49:37,596 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: Master,sub=Coprocessor.Master.CP_org.apache.hadoop.hbase.security.access.SecureTestUtil$MasterSyncObserver 2025-01-08T10:49:37,596 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T10:49:37,597 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2025-01-08T10:49:37,597 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testExportExpiredSnapshot 2025-01-08T10:49:41,723 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T10:50:05,715 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T10:50:08,499 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster-MemStoreChunkPool Statistics {}] regionserver.ChunkCreator$MemStoreChunkPool$StatisticsThread(417): data stats (chunk size=2097152): current pool size=7, created chunk count=10, reused chunk count=22, reuseRatio=68.75% 2025-01-08T10:50:08,503 DEBUG [master/67f42ee69b4e:0:becomeActiveMaster-MemStoreChunkPool Statistics {}] regionserver.ChunkCreator$MemStoreChunkPool$StatisticsThread(417): index stats (chunk size=209715): current pool size=0, created chunk count=0, reused chunk count=0, reuseRatio=0 2025-01-08T10:50:15,946 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T10:50:35,715 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. Process Thread Dump: Automatic Stack Trace every 60 seconds waiting on M:0;67f42ee69b4e:34551 224 active threads Thread 1 (main): State: TIMED_WAITING Blocked count: 2 Waited count: 4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.FutureTask.awaitDone(FutureTask.java:444) java.base@17.0.11/java.util.concurrent.FutureTask.get(FutureTask.java:203) app//org.junit.internal.runners.statements.FailOnTimeout.getResult(FailOnTimeout.java:167) app//org.junit.internal.runners.statements.FailOnTimeout.evaluate(FailOnTimeout.java:128) app//org.apache.hadoop.hbase.SystemExitRule$1.evaluate(SystemExitRule.java:39) app//org.junit.rules.RunRules.evaluate(RunRules.java:20) app//org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) app//org.junit.runners.ParentRunner.run(ParentRunner.java:413) app//org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:316) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:240) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:214) app//org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:155) app//org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:385) app//org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) app//org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:507) app//org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:495) Thread 2 (Reference Handler): State: RUNNABLE Blocked count: 3 Waited count: 0 Stack: java.base@17.0.11/java.lang.ref.Reference.waitForReferencePendingList(Native Method) java.base@17.0.11/java.lang.ref.Reference.processPendingReferences(Reference.java:253) java.base@17.0.11/java.lang.ref.Reference$ReferenceHandler.run(Reference.java:215) Thread 3 (Finalizer): State: WAITING Blocked count: 40 Waited count: 12 Waiting on java.lang.ref.ReferenceQueue$Lock@199c2830 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) java.base@17.0.11/java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:172) Thread 4 (Signal Dispatcher): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 12 (Common-Cleaner): State: TIMED_WAITING Blocked count: 18 Waited count: 16 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/jdk.internal.ref.CleanerImpl.run(CleanerImpl.java:140) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) java.base@17.0.11/jdk.internal.misc.InnocuousThread.run(InnocuousThread.java:162) Thread 13 (Notification Thread): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 14 (pool-1-thread-1): State: RUNNABLE Blocked count: 0 Waited count: 19 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.poll(EPollPort.java:200) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:281) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 15 (pool-1-thread-2): State: WAITING Blocked count: 0 Waited count: 17 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@31a5b7ac Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:275) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 16 (surefire-forkedjvm-stream-flusher): State: TIMED_WAITING Blocked count: 0 Waited count: 3378 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 18 (surefire-forkedjvm-command-thread): State: WAITING Blocked count: 0 Waited count: 35 Waiting on java.util.concurrent.CountDownLatch$Sync@12da1e3f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.CountDownLatch.await(CountDownLatch.java:230) java.base@17.0.11/sun.nio.ch.PendingFuture.get(PendingFuture.java:178) app//org.apache.maven.surefire.api.util.internal.Channels$2.read(Channels.java:127) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:284) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) app//org.apache.maven.surefire.api.util.internal.Channels$3.readImpl(Channels.java:169) app//org.apache.maven.surefire.api.util.internal.AbstractNoninterruptibleReadableChannel.read(AbstractNoninterruptibleReadableChannel.java:50) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:430) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:419) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.readMessageType(AbstractStreamDecoder.java:116) app//org.apache.maven.surefire.booter.stream.CommandDecoder.decode(CommandDecoder.java:77) app//org.apache.maven.surefire.booter.spi.CommandChannelDecoder.decode(CommandChannelDecoder.java:60) app//org.apache.maven.surefire.booter.CommandReader$CommandRunnable.run(CommandReader.java:290) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 22 (Time-limited test): State: RUNNABLE Blocked count: 12870 Waited count: 13417 Stack: java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo1(Native Method) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:197) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:154) app//org.apache.hadoop.hbase.util.ReflectionUtils.printThreadInfo(ReflectionUtils.java:181) app//org.apache.hadoop.hbase.util.Threads.printThreadInfo(Threads.java:186) app//org.apache.hadoop.hbase.util.Threads.threadDumpingIsAlive(Threads.java:113) app//org.apache.hadoop.hbase.LocalHBaseCluster.join(LocalHBaseCluster.java:394) app//org.apache.hadoop.hbase.MiniHBaseCluster.waitUntilShutDown(MiniHBaseCluster.java:921) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniHBaseCluster(HBaseTestingUtility.java:1359) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniCluster(HBaseTestingUtility.java:1341) app//org.apache.hadoop.hbase.snapshot.TestExportSnapshot.tearDownAfterClass(TestExportSnapshot.java:121) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) java.base@17.0.11/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) java.base@17.0.11/java.lang.reflect.Method.invoke(Method.java:568) app//org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) app//org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) app//org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) app//org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) app//org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) Thread 23 (org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner): State: WAITING Blocked count: 12 Waited count: 13 Waiting on java.lang.ref.ReferenceQueue$Lock@2fdea105 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) app//org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner.run(FileSystem.java:4171) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 25 (SSL Certificates Store Monitor): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.TaskQueue@33da7b85 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 34 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@44d16179): State: TIMED_WAITING Blocked count: 0 Waited count: 669 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 35 (GcTimeMonitor obsWindow = 60000, sleepInterval = 5000, maxGcTimePerc = 100): State: TIMED_WAITING Blocked count: 0 Waited count: 67 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.GcTimeMonitor.run(GcTimeMonitor.java:161) Thread 36 (pool-6-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 37 (qtp123845481-37): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 38 (qtp123845481-38): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 39 (qtp123845481-39): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 40 (qtp123845481-40): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 41 (qtp123845481-41-acceptor-0@781b1e42-ServerConnector@67ec1049{HTTP/1.1, (http/1.1)}{localhost:42647}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 42 (qtp123845481-42): State: TIMED_WAITING Blocked count: 0 Waited count: 6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 43 (qtp123845481-43): State: TIMED_WAITING Blocked count: 0 Waited count: 6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 44 (qtp123845481-44): State: TIMED_WAITING Blocked count: 0 Waited count: 6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 45 (Session-HouseKeeper-604db391-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 46 (pool-7-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 53 (FSEditLogAsync): State: WAITING Blocked count: 28 Waited count: 3355 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@60a45dc1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.dequeueEdit(FSEditLogAsync.java:241) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.run(FSEditLogAsync.java:250) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 55 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 56 (IPC Server idle connection scanner for port 37495): State: TIMED_WAITING Blocked count: 1 Waited count: 35 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 58 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 67 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 61 (org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor@33e58e69): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor.run(PendingReconstructionBlocks.java:267) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 62 (DatanodeAdminMonitor-0): State: TIMED_WAITING Blocked count: 0 Waited count: 112 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 49 (org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor@989c5e0): State: TIMED_WAITING Blocked count: 0 Waited count: 67 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor.run(HeartbeatManager.java:563) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 47 (RedundancyMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 116 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) java.base@17.0.11/java.lang.Thread.sleep(Thread.java:344) java.base@17.0.11/java.util.concurrent.TimeUnit.sleep(TimeUnit.java:446) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$RedundancyMonitor.run(BlockManager.java:5352) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 48 (MarkedDeleteBlockScrubberThread): State: TIMED_WAITING Blocked count: 0 Waited count: 32952 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$MarkedDeleteBlockScrubber.run(BlockManager.java:5326) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 51 (Block report processor): State: WAITING Blocked count: 1 Waited count: 1625 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@468106b7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.processQueue(BlockManager.java:5627) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.run(BlockManager.java:5614) Thread 57 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 54 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 64 (IPC Server handler 0 on default port 37495): State: TIMED_WAITING Blocked count: 70 Waited count: 2320 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 65 (IPC Server handler 1 on default port 37495): State: TIMED_WAITING Blocked count: 70 Waited count: 2319 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 66 (IPC Server handler 2 on default port 37495): State: TIMED_WAITING Blocked count: 66 Waited count: 2322 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 67 (IPC Server handler 3 on default port 37495): State: TIMED_WAITING Blocked count: 66 Waited count: 2318 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 68 (IPC Server handler 4 on default port 37495): State: TIMED_WAITING Blocked count: 82 Waited count: 2306 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 69 (pool-12-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 71 (org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor@7d0a4f0): State: TIMED_WAITING Blocked count: 0 Waited count: 167 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor.run(LeaseManager.java:537) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 72 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor@108b955): State: TIMED_WAITING Blocked count: 0 Waited count: 67 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor.run(FSNamesystem.java:4550) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 73 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller@2ea8d10a): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller.run(FSNamesystem.java:4592) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 74 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber@5985a9b3): State: TIMED_WAITING Blocked count: 0 Waited count: 4 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber.run(FSNamesystem.java:4689) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 75 (CacheReplicationMonitor(822228812)): State: TIMED_WAITING Blocked count: 0 Waited count: 13 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1759) app//org.apache.hadoop.hdfs.server.blockmanagement.CacheReplicationMonitor.run(CacheReplicationMonitor.java:186) Thread 85 (pool-18-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 86 (qtp1135609267-86): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 87 (qtp1135609267-87-acceptor-0@473ea113-ServerConnector@3c5cb06a{HTTP/1.1, (http/1.1)}{localhost:34891}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 88 (qtp1135609267-88): State: TIMED_WAITING Blocked count: 0 Waited count: 6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 89 (qtp1135609267-89): State: TIMED_WAITING Blocked count: 0 Waited count: 6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 90 (Session-HouseKeeper-4ee4d1fd-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 91 (nioEventLoopGroup-2-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 92 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@3900f107): State: TIMED_WAITING Blocked count: 0 Waited count: 665 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 94 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 95 (IPC Server idle connection scanner for port 45989): State: TIMED_WAITING Blocked count: 1 Waited count: 35 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 97 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 67 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 100 (Command processor): State: WAITING Blocked count: 3 Waited count: 256 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@18778abc Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 101 (BP-22138141-172.17.0.2-1736333101274 heartbeating to localhost/127.0.0.1:37495): State: TIMED_WAITING Blocked count: 1338 Waited count: 1456 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 102 (pool-20-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 84 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@8acd5ca): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 96 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 93 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 103 (IPC Server handler 0 on default port 45989): State: TIMED_WAITING Blocked count: 0 Waited count: 333 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 104 (IPC Server handler 1 on default port 45989): State: TIMED_WAITING Blocked count: 0 Waited count: 333 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 105 (IPC Server handler 2 on default port 45989): State: TIMED_WAITING Blocked count: 0 Waited count: 333 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 106 (IPC Server handler 3 on default port 45989): State: TIMED_WAITING Blocked count: 0 Waited count: 333 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 109 (IPC Server handler 4 on default port 45989): State: TIMED_WAITING Blocked count: 0 Waited count: 333 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 107 (IPC Client (175901922) connection to localhost/127.0.0.1:37495 from jenkins): State: TIMED_WAITING Blocked count: 1305 Waited count: 1306 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Thread 108 (IPC Parameter Sending Thread for localhost/127.0.0.1:37495): State: TIMED_WAITING Blocked count: 0 Waited count: 2060 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 119 (pool-26-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 120 (qtp40190603-120): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 121 (qtp40190603-121-acceptor-0@733970c9-ServerConnector@27b3056c{HTTP/1.1, (http/1.1)}{localhost:42931}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 122 (qtp40190603-122): State: TIMED_WAITING Blocked count: 0 Waited count: 6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 123 (qtp40190603-123): State: TIMED_WAITING Blocked count: 0 Waited count: 6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 124 (Session-HouseKeeper-66a6431-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 127 (nioEventLoopGroup-4-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 128 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@63d3adf1): State: TIMED_WAITING Blocked count: 0 Waited count: 664 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 130 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 131 (IPC Server idle connection scanner for port 37919): State: TIMED_WAITING Blocked count: 1 Waited count: 35 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 133 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 67 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 136 (Command processor): State: WAITING Blocked count: 0 Waited count: 237 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@262fd62 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 137 (BP-22138141-172.17.0.2-1736333101274 heartbeating to localhost/127.0.0.1:37495): State: TIMED_WAITING Blocked count: 1359 Waited count: 1458 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 138 (pool-29-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 118 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@3e496c77): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 132 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 129 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 139 (IPC Server handler 0 on default port 37919): State: TIMED_WAITING Blocked count: 0 Waited count: 332 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 140 (IPC Server handler 1 on default port 37919): State: TIMED_WAITING Blocked count: 0 Waited count: 332 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 141 (IPC Server handler 2 on default port 37919): State: TIMED_WAITING Blocked count: 0 Waited count: 332 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 142 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data1)): State: TIMED_WAITING Blocked count: 7 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 144 (IPC Server handler 3 on default port 37919): State: TIMED_WAITING Blocked count: 0 Waited count: 332 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 146 (IPC Server handler 4 on default port 37919): State: TIMED_WAITING Blocked count: 0 Waited count: 332 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 145 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data2)): State: TIMED_WAITING Blocked count: 10 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 163 (pool-40-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 158 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data1/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 159 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data3)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 164 (qtp1264850999-164): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 165 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data4)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 166 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data2/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 167 (qtp1264850999-167-acceptor-0@758b1918-ServerConnector@3362f87c{HTTP/1.1, (http/1.1)}{localhost:40601}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 171 (qtp1264850999-171): State: TIMED_WAITING Blocked count: 0 Waited count: 6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 172 (qtp1264850999-172): State: TIMED_WAITING Blocked count: 0 Waited count: 6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 176 (Session-HouseKeeper-26f1d292-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 180 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data3/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 181 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data4/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 183 (pool-15-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 186 (java.util.concurrent.ThreadPoolExecutor$Worker@2c953ef0[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 190 (pool-23-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 193 (java.util.concurrent.ThreadPoolExecutor$Worker@1c56948b[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 194 (nioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 195 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@6eba8751): State: TIMED_WAITING Blocked count: 0 Waited count: 663 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 197 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 198 (IPC Server idle connection scanner for port 45913): State: TIMED_WAITING Blocked count: 1 Waited count: 35 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 200 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 67 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 203 (Command processor): State: WAITING Blocked count: 1 Waited count: 275 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@336893f4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 204 (BP-22138141-172.17.0.2-1736333101274 heartbeating to localhost/127.0.0.1:37495): State: TIMED_WAITING Blocked count: 1325 Waited count: 1451 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 205 (pool-46-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 162 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@7b35f50f): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 199 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 196 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 206 (IPC Server handler 0 on default port 45913): State: TIMED_WAITING Blocked count: 0 Waited count: 332 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 207 (IPC Server handler 1 on default port 45913): State: TIMED_WAITING Blocked count: 0 Waited count: 332 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 208 (IPC Server handler 2 on default port 45913): State: TIMED_WAITING Blocked count: 0 Waited count: 332 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 209 (IPC Server handler 3 on default port 45913): State: TIMED_WAITING Blocked count: 0 Waited count: 332 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 210 (IPC Server handler 4 on default port 45913): State: TIMED_WAITING Blocked count: 0 Waited count: 332 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 213 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data5)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 214 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data6)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 220 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data5/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 221 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data6/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 225 (pool-37-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 228 (java.util.concurrent.ThreadPoolExecutor$Worker@5fd6a7f8[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 233 (FsDatasetAsyncDiskServiceFixer): State: TIMED_WAITING Blocked count: 0 Waited count: 12 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer.run(HBaseTestingUtility.java:599) Thread 236 (NIOServerCxnFactory.SelectorThread-1): State: RUNNABLE Blocked count: 2 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 235 (NIOServerCxnFactory.SelectorThread-0): State: RUNNABLE Blocked count: 16 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 237 (NIOServerCxnFactory.AcceptThread:localhost/127.0.0.1:52367): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.select(NIOServerCnxnFactory.java:205) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.run(NIOServerCnxnFactory.java:181) Thread 234 (ConnnectionExpirer): State: TIMED_WAITING Blocked count: 0 Waited count: 34 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.NIOServerCnxnFactory$ConnectionExpirerThread.run(NIOServerCnxnFactory.java:554) Thread 238 (SessionTracker): State: TIMED_WAITING Blocked count: 0 Waited count: 166 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.SessionTrackerImpl.run(SessionTrackerImpl.java:163) Thread 239 (SyncThread:0): State: WAITING Blocked count: 33 Waited count: 710 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4dc1adaf Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.SyncRequestProcessor.run(SyncRequestProcessor.java:170) Thread 240 (ProcessThread(sid:0 cport:52367):): State: WAITING Blocked count: 0 Waited count: 800 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@d2bfd92 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.PrepRequestProcessor.run(PrepRequestProcessor.java:142) Thread 241 (RequestThrottler): State: WAITING Blocked count: 0 Waited count: 833 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@3c783c72 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.RequestThrottler.run(RequestThrottler.java:147) Thread 242 (NIOWorkerThread-1): State: WAITING Blocked count: 0 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 244 (LeaseRenewer:jenkins@localhost:37495): State: TIMED_WAITING Blocked count: 9 Waited count: 342 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 253 (Time-limited test.named-queue-events-pool-0): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@ff82e77 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) app//com.lmax.disruptor.BlockingWaitStrategy.waitFor(BlockingWaitStrategy.java:47) app//com.lmax.disruptor.ProcessingSequenceBarrier.waitFor(ProcessingSequenceBarrier.java:56) app//com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:159) app//com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 254 (HBase-Metrics2-1): State: TIMED_WAITING Blocked count: 0 Waited count: 264 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 255 (RS-EventLoopGroup-1-1): State: RUNNABLE Blocked count: 50 Waited count: 1 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 256 (Time-limited test-SendThread(127.0.0.1:52367)): State: RUNNABLE Blocked count: 20 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:332) app//org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1289) Thread 257 (Time-limited test-EventThread): State: WAITING Blocked count: 2 Waited count: 58 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@200ba914 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:550) Thread 258 (NIOWorkerThread-2): State: WAITING Blocked count: 3 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 259 (NIOWorkerThread-3): State: WAITING Blocked count: 3 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 260 (NIOWorkerThread-4): State: WAITING Blocked count: 1 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 261 (NIOWorkerThread-5): State: WAITING Blocked count: 3 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 262 (zk-event-processor-pool-0): State: WAITING Blocked count: 35 Waited count: 87 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4d05f8cc Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 263 (NIOWorkerThread-6): State: WAITING Blocked count: 4 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 264 (NIOWorkerThread-7): State: WAITING Blocked count: 3 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 265 (NIOWorkerThread-8): State: WAITING Blocked count: 0 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 266 (NIOWorkerThread-9): State: WAITING Blocked count: 2 Waited count: 131 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 267 (NIOWorkerThread-10): State: WAITING Blocked count: 3 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 268 (NIOWorkerThread-11): State: WAITING Blocked count: 6 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 269 (NIOWorkerThread-12): State: WAITING Blocked count: 1 Waited count: 131 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 270 (NIOWorkerThread-13): State: WAITING Blocked count: 2 Waited count: 131 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 271 (NIOWorkerThread-14): State: WAITING Blocked count: 3 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 272 (NIOWorkerThread-15): State: WAITING Blocked count: 2 Waited count: 131 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 273 (NIOWorkerThread-16): State: WAITING Blocked count: 2 Waited count: 131 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 274 (RpcServer.default.FPBQ.Fifo.handler=0,queue=0,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@3c0ec8d Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 275 (RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551): State: WAITING Blocked count: 221 Waited count: 834 Waiting on java.util.concurrent.Semaphore$NonfairSync@32b787af Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 276 (RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551): State: WAITING Blocked count: 93 Waited count: 356 Waiting on java.util.concurrent.Semaphore$NonfairSync@4f89cbe4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 277 (RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=34551): State: WAITING Blocked count: 80 Waited count: 5313 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@2d289941 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 278 (RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@2988bced Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 279 (RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@2988bced Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 280 (RpcServer.replication.FPBQ.Fifo.handler=0,queue=0,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@278ccc8c Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 281 (RpcServer.replication.FPBQ.Fifo.handler=1,queue=0,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@265be572 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 282 (RpcServer.replication.FPBQ.Fifo.handler=2,queue=0,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@4fd08269 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 283 (RpcServer.metaPriority.FPBQ.Fifo.handler=0,queue=0,port=34551): State: WAITING Blocked count: 0 Waited count: 3 Waiting on java.util.concurrent.Semaphore$NonfairSync@23d320aa Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 287 (RS-EventLoopGroup-3-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 307 (RS-EventLoopGroup-4-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 327 (RS-EventLoopGroup-5-1): State: RUNNABLE Blocked count: 90 Waited count: 4 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 285 (M:0;67f42ee69b4e:34551): State: TIMED_WAITING Blocked count: 6 Waited count: 2682 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1759) app//org.apache.hadoop.hbase.regionserver.wal.SyncFuture.get(SyncFuture.java:169) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.blockOnSync(AbstractFSWAL.java:879) app//org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.doSync(AsyncFSWAL.java:671) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.lambda$sync$1(AbstractFSWAL.java:600) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL$$Lambda$983/0x00007f478cf19800.run(Unknown Source) app//org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:600) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:590) app//org.apache.hadoop.hbase.regionserver.HRegion.doSyncOfUnflushedWALChanges(HRegion.java:2869) app//org.apache.hadoop.hbase.regionserver.HRegion.internalPrepareFlushCache(HRegion.java:2811) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2670) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2644) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2635) app//org.apache.hadoop.hbase.regionserver.HRegion.doClose(HRegion.java:1810) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1631) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1586) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1569) Thread 348 (Monitor thread for TaskMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 33 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.monitoring.TaskMonitor$MonitorRunnable.run(TaskMonitor.java:325) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 350 (master/67f42ee69b4e:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 352 (master/67f42ee69b4e:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 354 (org.apache.hadoop.hdfs.PeerCache@58490a7b): State: TIMED_WAITING Blocked count: 0 Waited count: 110 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.PeerCache.run(PeerCache.java:253) app//org.apache.hadoop.hdfs.PeerCache.access$000(PeerCache.java:46) app//org.apache.hadoop.hdfs.PeerCache$1.run(PeerCache.java:124) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 373 (master:store-WAL-Roller): State: TIMED_WAITING Blocked count: 0 Waited count: 3274 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hbase.wal.AbstractWALRoller.run(AbstractWALRoller.java:179) Thread 390 (RS-EventLoopGroup-5-2): State: RUNNABLE Blocked count: 108 Waited count: 3 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 391 (RS-EventLoopGroup-5-3): State: RUNNABLE Blocked count: 131 Waited count: 3 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 404 (Idle-Rpc-Conn-Sweeper-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 53 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 415 (SnapshotHandlerChoreCleaner): State: TIMED_WAITING Blocked count: 0 Waited count: 33 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 403 (RpcClient-timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 32680 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 424 (RS-EventLoopGroup-1-2): State: RUNNABLE Blocked count: 27 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 425 (RS-EventLoopGroup-1-3): State: RUNNABLE Blocked count: 56 Waited count: 2 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 450 (RegionServerTracker-0): State: WAITING Blocked count: 8 Waited count: 10 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5be54308 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 472 (regionserver/67f42ee69b4e:0.procedureResultReporter): State: WAITING Blocked count: 14 Waited count: 29 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@508e73da Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 473 (regionserver/67f42ee69b4e:0.procedureResultReporter): State: WAITING Blocked count: 22 Waited count: 33 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@430dcb4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 474 (regionserver/67f42ee69b4e:0.procedureResultReporter): State: WAITING Blocked count: 14 Waited count: 23 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@595b70f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 501 (LeaseRenewer:jenkins.hfs.1@localhost:37495): State: TIMED_WAITING Blocked count: 8 Waited count: 340 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 504 (LeaseRenewer:jenkins.hfs.2@localhost:37495): State: TIMED_WAITING Blocked count: 8 Waited count: 339 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 507 (LeaseRenewer:jenkins.hfs.0@localhost:37495): State: TIMED_WAITING Blocked count: 8 Waited count: 339 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 512 (RS-EventLoopGroup-4-2): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 523 (RS-EventLoopGroup-4-3): State: RUNNABLE Blocked count: 2 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 525 (region-location-0): State: WAITING Blocked count: 8 Waited count: 12 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@741949b6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 551 (Async-Client-Retry-Timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 32454 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 566 (RPCClient-NioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 3 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 583 (ForkJoinPool.commonPool-worker-2): State: WAITING Blocked count: 0 Waited count: 541 Waiting on java.util.concurrent.ForkJoinPool@57aaf44c Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1724) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Thread 584 (ForkJoinPool.commonPool-worker-3): State: TIMED_WAITING Blocked count: 0 Waited count: 713 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkUntil(LockSupport.java:410) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1726) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Thread 606 (region-location-1): State: WAITING Blocked count: 3 Waited count: 7 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@741949b6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 607 (region-location-2): State: WAITING Blocked count: 3 Waited count: 6 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@741949b6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 608 (region-location-3): State: WAITING Blocked count: 2 Waited count: 5 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@741949b6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1012 (MutableQuantiles-0): State: TIMED_WAITING Blocked count: 0 Waited count: 359 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1073 (RPCClient-NioEventLoopGroup-6-2): State: RUNNABLE Blocked count: 6 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1101 (RS-EventLoopGroup-3-2): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1112 (zk-permission-watcher-pool-0): State: WAITING Blocked count: 62 Waited count: 97 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7d78af0e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1120 (RPCClient-NioEventLoopGroup-6-3): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1121 (RS-EventLoopGroup-3-3): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1562 (Container metrics unregistration): State: WAITING Blocked count: 10 Waited count: 33 Waiting on java.util.TaskQueue@43921dee Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 2944 (RPCClient-NioEventLoopGroup-6-4): State: RUNNABLE Blocked count: 2 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 3542 (region-location-4): State: WAITING Blocked count: 3 Waited count: 6 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@741949b6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5203 (RPCClient-NioEventLoopGroup-6-5): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5204 (RPCClient-NioEventLoopGroup-6-6): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5205 (RPCClient-NioEventLoopGroup-6-7): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9395 (AsyncFSWAL-1-hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/MasterData-prefix:67f42ee69b4e,34551,1736333106957): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@431a8944 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9399 (Timer for 'JobHistoryServer' metrics system): State: TIMED_WAITING Blocked count: 0 Waited count: 3 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) 2025-01-08T10:51:05,716 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T10:51:35,716 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. Process Thread Dump: Automatic Stack Trace every 60 seconds waiting on M:0;67f42ee69b4e:34551 219 active threads Thread 1 (main): State: TIMED_WAITING Blocked count: 2 Waited count: 4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.FutureTask.awaitDone(FutureTask.java:444) java.base@17.0.11/java.util.concurrent.FutureTask.get(FutureTask.java:203) app//org.junit.internal.runners.statements.FailOnTimeout.getResult(FailOnTimeout.java:167) app//org.junit.internal.runners.statements.FailOnTimeout.evaluate(FailOnTimeout.java:128) app//org.apache.hadoop.hbase.SystemExitRule$1.evaluate(SystemExitRule.java:39) app//org.junit.rules.RunRules.evaluate(RunRules.java:20) app//org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) app//org.junit.runners.ParentRunner.run(ParentRunner.java:413) app//org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:316) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:240) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:214) app//org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:155) app//org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:385) app//org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) app//org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:507) app//org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:495) Thread 2 (Reference Handler): State: RUNNABLE Blocked count: 3 Waited count: 0 Stack: java.base@17.0.11/java.lang.ref.Reference.waitForReferencePendingList(Native Method) java.base@17.0.11/java.lang.ref.Reference.processPendingReferences(Reference.java:253) java.base@17.0.11/java.lang.ref.Reference$ReferenceHandler.run(Reference.java:215) Thread 3 (Finalizer): State: WAITING Blocked count: 40 Waited count: 12 Waiting on java.lang.ref.ReferenceQueue$Lock@199c2830 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) java.base@17.0.11/java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:172) Thread 4 (Signal Dispatcher): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 12 (Common-Cleaner): State: TIMED_WAITING Blocked count: 18 Waited count: 17 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/jdk.internal.ref.CleanerImpl.run(CleanerImpl.java:140) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) java.base@17.0.11/jdk.internal.misc.InnocuousThread.run(InnocuousThread.java:162) Thread 13 (Notification Thread): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 14 (pool-1-thread-1): State: RUNNABLE Blocked count: 0 Waited count: 22 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.poll(EPollPort.java:200) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:281) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 15 (pool-1-thread-2): State: WAITING Blocked count: 0 Waited count: 20 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@31a5b7ac Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:275) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 16 (surefire-forkedjvm-stream-flusher): State: TIMED_WAITING Blocked count: 0 Waited count: 3978 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 18 (surefire-forkedjvm-command-thread): State: WAITING Blocked count: 0 Waited count: 41 Waiting on java.util.concurrent.CountDownLatch$Sync@445b48c4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.CountDownLatch.await(CountDownLatch.java:230) java.base@17.0.11/sun.nio.ch.PendingFuture.get(PendingFuture.java:178) app//org.apache.maven.surefire.api.util.internal.Channels$2.read(Channels.java:127) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:284) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) app//org.apache.maven.surefire.api.util.internal.Channels$3.readImpl(Channels.java:169) app//org.apache.maven.surefire.api.util.internal.AbstractNoninterruptibleReadableChannel.read(AbstractNoninterruptibleReadableChannel.java:50) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:430) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:419) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.readMessageType(AbstractStreamDecoder.java:116) app//org.apache.maven.surefire.booter.stream.CommandDecoder.decode(CommandDecoder.java:77) app//org.apache.maven.surefire.booter.spi.CommandChannelDecoder.decode(CommandChannelDecoder.java:60) app//org.apache.maven.surefire.booter.CommandReader$CommandRunnable.run(CommandReader.java:290) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 22 (Time-limited test): State: RUNNABLE Blocked count: 12870 Waited count: 13418 Stack: java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo1(Native Method) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:197) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:154) app//org.apache.hadoop.hbase.util.ReflectionUtils.printThreadInfo(ReflectionUtils.java:181) app//org.apache.hadoop.hbase.util.Threads.printThreadInfo(Threads.java:186) app//org.apache.hadoop.hbase.util.Threads.threadDumpingIsAlive(Threads.java:113) app//org.apache.hadoop.hbase.LocalHBaseCluster.join(LocalHBaseCluster.java:394) app//org.apache.hadoop.hbase.MiniHBaseCluster.waitUntilShutDown(MiniHBaseCluster.java:921) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniHBaseCluster(HBaseTestingUtility.java:1359) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniCluster(HBaseTestingUtility.java:1341) app//org.apache.hadoop.hbase.snapshot.TestExportSnapshot.tearDownAfterClass(TestExportSnapshot.java:121) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) java.base@17.0.11/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) java.base@17.0.11/java.lang.reflect.Method.invoke(Method.java:568) app//org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) app//org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) app//org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) app//org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) app//org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) Thread 23 (org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner): State: WAITING Blocked count: 12 Waited count: 13 Waiting on java.lang.ref.ReferenceQueue$Lock@2fdea105 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) app//org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner.run(FileSystem.java:4171) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 25 (SSL Certificates Store Monitor): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.TaskQueue@33da7b85 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 34 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@44d16179): State: TIMED_WAITING Blocked count: 0 Waited count: 789 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 35 (GcTimeMonitor obsWindow = 60000, sleepInterval = 5000, maxGcTimePerc = 100): State: TIMED_WAITING Blocked count: 0 Waited count: 79 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.GcTimeMonitor.run(GcTimeMonitor.java:161) Thread 36 (pool-6-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 37 (qtp123845481-37): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 38 (qtp123845481-38): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 39 (qtp123845481-39): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 40 (qtp123845481-40): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 41 (qtp123845481-41-acceptor-0@781b1e42-ServerConnector@67ec1049{HTTP/1.1, (http/1.1)}{localhost:42647}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 42 (qtp123845481-42): State: TIMED_WAITING Blocked count: 0 Waited count: 7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 43 (qtp123845481-43): State: TIMED_WAITING Blocked count: 0 Waited count: 7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 44 (qtp123845481-44): State: TIMED_WAITING Blocked count: 0 Waited count: 7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 45 (Session-HouseKeeper-604db391-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 46 (pool-7-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 53 (FSEditLogAsync): State: WAITING Blocked count: 28 Waited count: 3355 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@60a45dc1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.dequeueEdit(FSEditLogAsync.java:241) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.run(FSEditLogAsync.java:250) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 55 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 56 (IPC Server idle connection scanner for port 37495): State: TIMED_WAITING Blocked count: 1 Waited count: 41 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 58 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 79 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 61 (org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor@33e58e69): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor.run(PendingReconstructionBlocks.java:267) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 62 (DatanodeAdminMonitor-0): State: TIMED_WAITING Blocked count: 0 Waited count: 132 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 49 (org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor@989c5e0): State: TIMED_WAITING Blocked count: 0 Waited count: 79 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor.run(HeartbeatManager.java:563) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 47 (RedundancyMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 136 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) java.base@17.0.11/java.lang.Thread.sleep(Thread.java:344) java.base@17.0.11/java.util.concurrent.TimeUnit.sleep(TimeUnit.java:446) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$RedundancyMonitor.run(BlockManager.java:5352) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 48 (MarkedDeleteBlockScrubberThread): State: TIMED_WAITING Blocked count: 0 Waited count: 38915 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$MarkedDeleteBlockScrubber.run(BlockManager.java:5326) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 51 (Block report processor): State: WAITING Blocked count: 1 Waited count: 1625 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@468106b7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.processQueue(BlockManager.java:5627) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.run(BlockManager.java:5614) Thread 57 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 54 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 64 (IPC Server handler 0 on default port 37495): State: TIMED_WAITING Blocked count: 70 Waited count: 2380 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 65 (IPC Server handler 1 on default port 37495): State: TIMED_WAITING Blocked count: 70 Waited count: 2380 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 66 (IPC Server handler 2 on default port 37495): State: TIMED_WAITING Blocked count: 66 Waited count: 2383 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 67 (IPC Server handler 3 on default port 37495): State: TIMED_WAITING Blocked count: 66 Waited count: 2379 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 68 (IPC Server handler 4 on default port 37495): State: TIMED_WAITING Blocked count: 82 Waited count: 2367 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 69 (pool-12-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 71 (org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor@7d0a4f0): State: TIMED_WAITING Blocked count: 0 Waited count: 197 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor.run(LeaseManager.java:537) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 72 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor@108b955): State: TIMED_WAITING Blocked count: 0 Waited count: 79 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor.run(FSNamesystem.java:4550) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 73 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller@2ea8d10a): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller.run(FSNamesystem.java:4592) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 74 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber@5985a9b3): State: TIMED_WAITING Blocked count: 0 Waited count: 4 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber.run(FSNamesystem.java:4689) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 75 (CacheReplicationMonitor(822228812)): State: TIMED_WAITING Blocked count: 0 Waited count: 15 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1759) app//org.apache.hadoop.hdfs.server.blockmanagement.CacheReplicationMonitor.run(CacheReplicationMonitor.java:186) Thread 85 (pool-18-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 86 (qtp1135609267-86): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 87 (qtp1135609267-87-acceptor-0@473ea113-ServerConnector@3c5cb06a{HTTP/1.1, (http/1.1)}{localhost:34891}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 88 (qtp1135609267-88): State: TIMED_WAITING Blocked count: 0 Waited count: 7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 89 (qtp1135609267-89): State: TIMED_WAITING Blocked count: 0 Waited count: 7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 90 (Session-HouseKeeper-4ee4d1fd-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 91 (nioEventLoopGroup-2-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 92 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@3900f107): State: TIMED_WAITING Blocked count: 0 Waited count: 786 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 94 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 95 (IPC Server idle connection scanner for port 45989): State: TIMED_WAITING Blocked count: 1 Waited count: 41 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 97 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 79 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 100 (Command processor): State: WAITING Blocked count: 3 Waited count: 276 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@18778abc Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 101 (BP-22138141-172.17.0.2-1736333101274 heartbeating to localhost/127.0.0.1:37495): State: TIMED_WAITING Blocked count: 1358 Waited count: 1496 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 102 (pool-20-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 84 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@8acd5ca): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 96 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 93 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 103 (IPC Server handler 0 on default port 45989): State: TIMED_WAITING Blocked count: 0 Waited count: 393 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 104 (IPC Server handler 1 on default port 45989): State: TIMED_WAITING Blocked count: 0 Waited count: 393 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 105 (IPC Server handler 2 on default port 45989): State: TIMED_WAITING Blocked count: 0 Waited count: 393 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 106 (IPC Server handler 3 on default port 45989): State: TIMED_WAITING Blocked count: 0 Waited count: 394 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 109 (IPC Server handler 4 on default port 45989): State: TIMED_WAITING Blocked count: 0 Waited count: 393 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 107 (IPC Client (175901922) connection to localhost/127.0.0.1:37495 from jenkins): State: TIMED_WAITING Blocked count: 1365 Waited count: 1366 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Thread 108 (IPC Parameter Sending Thread for localhost/127.0.0.1:37495): State: TIMED_WAITING Blocked count: 0 Waited count: 2120 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 119 (pool-26-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 120 (qtp40190603-120): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 121 (qtp40190603-121-acceptor-0@733970c9-ServerConnector@27b3056c{HTTP/1.1, (http/1.1)}{localhost:42931}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 122 (qtp40190603-122): State: TIMED_WAITING Blocked count: 0 Waited count: 7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 123 (qtp40190603-123): State: TIMED_WAITING Blocked count: 0 Waited count: 7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 124 (Session-HouseKeeper-66a6431-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 127 (nioEventLoopGroup-4-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 128 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@63d3adf1): State: TIMED_WAITING Blocked count: 0 Waited count: 784 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 130 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 131 (IPC Server idle connection scanner for port 37919): State: TIMED_WAITING Blocked count: 1 Waited count: 41 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 133 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 79 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 136 (Command processor): State: WAITING Blocked count: 0 Waited count: 257 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@262fd62 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 137 (BP-22138141-172.17.0.2-1736333101274 heartbeating to localhost/127.0.0.1:37495): State: TIMED_WAITING Blocked count: 1379 Waited count: 1498 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 138 (pool-29-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 118 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@3e496c77): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 132 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 129 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 139 (IPC Server handler 0 on default port 37919): State: TIMED_WAITING Blocked count: 0 Waited count: 392 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 140 (IPC Server handler 1 on default port 37919): State: TIMED_WAITING Blocked count: 0 Waited count: 392 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 141 (IPC Server handler 2 on default port 37919): State: TIMED_WAITING Blocked count: 0 Waited count: 392 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 142 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data1)): State: TIMED_WAITING Blocked count: 7 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 144 (IPC Server handler 3 on default port 37919): State: TIMED_WAITING Blocked count: 0 Waited count: 392 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 146 (IPC Server handler 4 on default port 37919): State: TIMED_WAITING Blocked count: 0 Waited count: 392 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 145 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data2)): State: TIMED_WAITING Blocked count: 10 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 163 (pool-40-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 158 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data1/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 159 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data3)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 164 (qtp1264850999-164): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 165 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data4)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 166 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data2/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 167 (qtp1264850999-167-acceptor-0@758b1918-ServerConnector@3362f87c{HTTP/1.1, (http/1.1)}{localhost:40601}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 171 (qtp1264850999-171): State: TIMED_WAITING Blocked count: 0 Waited count: 7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 172 (qtp1264850999-172): State: TIMED_WAITING Blocked count: 0 Waited count: 7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 176 (Session-HouseKeeper-26f1d292-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 180 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data3/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 181 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data4/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 183 (pool-15-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 186 (java.util.concurrent.ThreadPoolExecutor$Worker@2c953ef0[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 190 (pool-23-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 193 (java.util.concurrent.ThreadPoolExecutor$Worker@1c56948b[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 194 (nioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 195 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@6eba8751): State: TIMED_WAITING Blocked count: 0 Waited count: 783 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 197 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 198 (IPC Server idle connection scanner for port 45913): State: TIMED_WAITING Blocked count: 1 Waited count: 41 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 200 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 79 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 203 (Command processor): State: WAITING Blocked count: 1 Waited count: 295 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@336893f4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 204 (BP-22138141-172.17.0.2-1736333101274 heartbeating to localhost/127.0.0.1:37495): State: TIMED_WAITING Blocked count: 1345 Waited count: 1491 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 205 (pool-46-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 162 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@7b35f50f): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 199 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 196 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 206 (IPC Server handler 0 on default port 45913): State: TIMED_WAITING Blocked count: 0 Waited count: 392 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 207 (IPC Server handler 1 on default port 45913): State: TIMED_WAITING Blocked count: 0 Waited count: 392 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 208 (IPC Server handler 2 on default port 45913): State: TIMED_WAITING Blocked count: 0 Waited count: 392 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 209 (IPC Server handler 3 on default port 45913): State: TIMED_WAITING Blocked count: 0 Waited count: 392 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 210 (IPC Server handler 4 on default port 45913): State: TIMED_WAITING Blocked count: 0 Waited count: 392 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 213 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data5)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 214 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data6)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 220 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data5/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 221 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data6/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 225 (pool-37-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 228 (java.util.concurrent.ThreadPoolExecutor$Worker@5fd6a7f8[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 233 (FsDatasetAsyncDiskServiceFixer): State: TIMED_WAITING Blocked count: 0 Waited count: 14 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer.run(HBaseTestingUtility.java:599) Thread 236 (NIOServerCxnFactory.SelectorThread-1): State: RUNNABLE Blocked count: 2 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 235 (NIOServerCxnFactory.SelectorThread-0): State: RUNNABLE Blocked count: 16 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 237 (NIOServerCxnFactory.AcceptThread:localhost/127.0.0.1:52367): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.select(NIOServerCnxnFactory.java:205) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.run(NIOServerCnxnFactory.java:181) Thread 234 (ConnnectionExpirer): State: TIMED_WAITING Blocked count: 0 Waited count: 40 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.NIOServerCnxnFactory$ConnectionExpirerThread.run(NIOServerCnxnFactory.java:554) Thread 238 (SessionTracker): State: TIMED_WAITING Blocked count: 0 Waited count: 196 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.SessionTrackerImpl.run(SessionTrackerImpl.java:163) Thread 239 (SyncThread:0): State: WAITING Blocked count: 33 Waited count: 715 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4dc1adaf Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.SyncRequestProcessor.run(SyncRequestProcessor.java:170) Thread 240 (ProcessThread(sid:0 cport:52367):): State: WAITING Blocked count: 0 Waited count: 805 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@d2bfd92 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.PrepRequestProcessor.run(PrepRequestProcessor.java:142) Thread 241 (RequestThrottler): State: WAITING Blocked count: 0 Waited count: 838 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@3c783c72 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.RequestThrottler.run(RequestThrottler.java:147) Thread 242 (NIOWorkerThread-1): State: WAITING Blocked count: 0 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 253 (Time-limited test.named-queue-events-pool-0): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@ff82e77 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) app//com.lmax.disruptor.BlockingWaitStrategy.waitFor(BlockingWaitStrategy.java:47) app//com.lmax.disruptor.ProcessingSequenceBarrier.waitFor(ProcessingSequenceBarrier.java:56) app//com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:159) app//com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 254 (HBase-Metrics2-1): State: TIMED_WAITING Blocked count: 0 Waited count: 292 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 255 (RS-EventLoopGroup-1-1): State: RUNNABLE Blocked count: 50 Waited count: 1 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 256 (Time-limited test-SendThread(127.0.0.1:52367)): State: RUNNABLE Blocked count: 20 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:332) app//org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1289) Thread 257 (Time-limited test-EventThread): State: WAITING Blocked count: 2 Waited count: 58 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@200ba914 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:550) Thread 258 (NIOWorkerThread-2): State: WAITING Blocked count: 3 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 259 (NIOWorkerThread-3): State: WAITING Blocked count: 3 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 260 (NIOWorkerThread-4): State: WAITING Blocked count: 1 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 261 (NIOWorkerThread-5): State: WAITING Blocked count: 3 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 262 (zk-event-processor-pool-0): State: WAITING Blocked count: 35 Waited count: 87 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4d05f8cc Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 263 (NIOWorkerThread-6): State: WAITING Blocked count: 4 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 264 (NIOWorkerThread-7): State: WAITING Blocked count: 3 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 265 (NIOWorkerThread-8): State: WAITING Blocked count: 0 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 266 (NIOWorkerThread-9): State: WAITING Blocked count: 2 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 267 (NIOWorkerThread-10): State: WAITING Blocked count: 3 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 268 (NIOWorkerThread-11): State: WAITING Blocked count: 6 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 269 (NIOWorkerThread-12): State: WAITING Blocked count: 1 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 270 (NIOWorkerThread-13): State: WAITING Blocked count: 2 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 271 (NIOWorkerThread-14): State: WAITING Blocked count: 3 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 272 (NIOWorkerThread-15): State: WAITING Blocked count: 2 Waited count: 131 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 273 (NIOWorkerThread-16): State: WAITING Blocked count: 2 Waited count: 131 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 274 (RpcServer.default.FPBQ.Fifo.handler=0,queue=0,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@3c0ec8d Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 275 (RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551): State: WAITING Blocked count: 221 Waited count: 834 Waiting on java.util.concurrent.Semaphore$NonfairSync@32b787af Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 276 (RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551): State: WAITING Blocked count: 93 Waited count: 356 Waiting on java.util.concurrent.Semaphore$NonfairSync@4f89cbe4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 277 (RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=34551): State: WAITING Blocked count: 80 Waited count: 5313 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@2d289941 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 278 (RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@2988bced Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 279 (RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@2988bced Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 280 (RpcServer.replication.FPBQ.Fifo.handler=0,queue=0,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@278ccc8c Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 281 (RpcServer.replication.FPBQ.Fifo.handler=1,queue=0,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@265be572 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 282 (RpcServer.replication.FPBQ.Fifo.handler=2,queue=0,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@4fd08269 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 283 (RpcServer.metaPriority.FPBQ.Fifo.handler=0,queue=0,port=34551): State: WAITING Blocked count: 0 Waited count: 3 Waiting on java.util.concurrent.Semaphore$NonfairSync@23d320aa Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 287 (RS-EventLoopGroup-3-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 307 (RS-EventLoopGroup-4-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 327 (RS-EventLoopGroup-5-1): State: RUNNABLE Blocked count: 90 Waited count: 4 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 285 (M:0;67f42ee69b4e:34551): State: TIMED_WAITING Blocked count: 6 Waited count: 2682 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1759) app//org.apache.hadoop.hbase.regionserver.wal.SyncFuture.get(SyncFuture.java:169) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.blockOnSync(AbstractFSWAL.java:879) app//org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.doSync(AsyncFSWAL.java:671) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.lambda$sync$1(AbstractFSWAL.java:600) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL$$Lambda$983/0x00007f478cf19800.run(Unknown Source) app//org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:600) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:590) app//org.apache.hadoop.hbase.regionserver.HRegion.doSyncOfUnflushedWALChanges(HRegion.java:2869) app//org.apache.hadoop.hbase.regionserver.HRegion.internalPrepareFlushCache(HRegion.java:2811) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2670) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2644) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2635) app//org.apache.hadoop.hbase.regionserver.HRegion.doClose(HRegion.java:1810) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1631) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1586) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1569) Thread 348 (Monitor thread for TaskMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 39 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.monitoring.TaskMonitor$MonitorRunnable.run(TaskMonitor.java:325) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 350 (master/67f42ee69b4e:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 352 (master/67f42ee69b4e:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 354 (org.apache.hadoop.hdfs.PeerCache@58490a7b): State: TIMED_WAITING Blocked count: 0 Waited count: 130 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.PeerCache.run(PeerCache.java:253) app//org.apache.hadoop.hdfs.PeerCache.access$000(PeerCache.java:46) app//org.apache.hadoop.hdfs.PeerCache$1.run(PeerCache.java:124) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 373 (master:store-WAL-Roller): State: TIMED_WAITING Blocked count: 0 Waited count: 3873 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hbase.wal.AbstractWALRoller.run(AbstractWALRoller.java:179) Thread 390 (RS-EventLoopGroup-5-2): State: RUNNABLE Blocked count: 108 Waited count: 3 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 391 (RS-EventLoopGroup-5-3): State: RUNNABLE Blocked count: 131 Waited count: 3 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 404 (Idle-Rpc-Conn-Sweeper-pool-0): State: WAITING Blocked count: 0 Waited count: 65 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6a97dfa0 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 415 (SnapshotHandlerChoreCleaner): State: TIMED_WAITING Blocked count: 0 Waited count: 39 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 403 (RpcClient-timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 38682 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 424 (RS-EventLoopGroup-1-2): State: RUNNABLE Blocked count: 27 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 425 (RS-EventLoopGroup-1-3): State: RUNNABLE Blocked count: 56 Waited count: 2 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 450 (RegionServerTracker-0): State: WAITING Blocked count: 8 Waited count: 10 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5be54308 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 472 (regionserver/67f42ee69b4e:0.procedureResultReporter): State: WAITING Blocked count: 14 Waited count: 29 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@508e73da Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 473 (regionserver/67f42ee69b4e:0.procedureResultReporter): State: WAITING Blocked count: 22 Waited count: 33 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@430dcb4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 474 (regionserver/67f42ee69b4e:0.procedureResultReporter): State: WAITING Blocked count: 14 Waited count: 23 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@595b70f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 512 (RS-EventLoopGroup-4-2): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 523 (RS-EventLoopGroup-4-3): State: RUNNABLE Blocked count: 2 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 525 (region-location-0): State: WAITING Blocked count: 8 Waited count: 12 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@741949b6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 551 (Async-Client-Retry-Timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 38456 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 566 (RPCClient-NioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 3 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 583 (ForkJoinPool.commonPool-worker-2): State: TIMED_WAITING Blocked count: 0 Waited count: 542 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkUntil(LockSupport.java:410) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1726) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Thread 606 (region-location-1): State: WAITING Blocked count: 3 Waited count: 7 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@741949b6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 607 (region-location-2): State: WAITING Blocked count: 3 Waited count: 6 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@741949b6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 608 (region-location-3): State: WAITING Blocked count: 2 Waited count: 5 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@741949b6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1012 (MutableQuantiles-0): State: TIMED_WAITING Blocked count: 0 Waited count: 365 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1073 (RPCClient-NioEventLoopGroup-6-2): State: RUNNABLE Blocked count: 6 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1101 (RS-EventLoopGroup-3-2): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1112 (zk-permission-watcher-pool-0): State: WAITING Blocked count: 62 Waited count: 97 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7d78af0e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1120 (RPCClient-NioEventLoopGroup-6-3): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1121 (RS-EventLoopGroup-3-3): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1562 (Container metrics unregistration): State: WAITING Blocked count: 10 Waited count: 33 Waiting on java.util.TaskQueue@43921dee Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 2944 (RPCClient-NioEventLoopGroup-6-4): State: RUNNABLE Blocked count: 2 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 3542 (region-location-4): State: WAITING Blocked count: 3 Waited count: 6 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@741949b6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5203 (RPCClient-NioEventLoopGroup-6-5): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5204 (RPCClient-NioEventLoopGroup-6-6): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5205 (RPCClient-NioEventLoopGroup-6-7): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9395 (AsyncFSWAL-1-hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/MasterData-prefix:67f42ee69b4e,34551,1736333106957): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@431a8944 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9399 (Timer for 'JobHistoryServer' metrics system): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) 2025-01-08T10:52:05,716 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T10:52:35,716 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. Process Thread Dump: Automatic Stack Trace every 60 seconds waiting on M:0;67f42ee69b4e:34551 218 active threads Thread 1 (main): State: TIMED_WAITING Blocked count: 2 Waited count: 4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.FutureTask.awaitDone(FutureTask.java:444) java.base@17.0.11/java.util.concurrent.FutureTask.get(FutureTask.java:203) app//org.junit.internal.runners.statements.FailOnTimeout.getResult(FailOnTimeout.java:167) app//org.junit.internal.runners.statements.FailOnTimeout.evaluate(FailOnTimeout.java:128) app//org.apache.hadoop.hbase.SystemExitRule$1.evaluate(SystemExitRule.java:39) app//org.junit.rules.RunRules.evaluate(RunRules.java:20) app//org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) app//org.junit.runners.ParentRunner.run(ParentRunner.java:413) app//org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:316) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:240) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:214) app//org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:155) app//org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:385) app//org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) app//org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:507) app//org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:495) Thread 2 (Reference Handler): State: RUNNABLE Blocked count: 3 Waited count: 0 Stack: java.base@17.0.11/java.lang.ref.Reference.waitForReferencePendingList(Native Method) java.base@17.0.11/java.lang.ref.Reference.processPendingReferences(Reference.java:253) java.base@17.0.11/java.lang.ref.Reference$ReferenceHandler.run(Reference.java:215) Thread 3 (Finalizer): State: WAITING Blocked count: 40 Waited count: 12 Waiting on java.lang.ref.ReferenceQueue$Lock@199c2830 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) java.base@17.0.11/java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:172) Thread 4 (Signal Dispatcher): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 12 (Common-Cleaner): State: TIMED_WAITING Blocked count: 18 Waited count: 18 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/jdk.internal.ref.CleanerImpl.run(CleanerImpl.java:140) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) java.base@17.0.11/jdk.internal.misc.InnocuousThread.run(InnocuousThread.java:162) Thread 13 (Notification Thread): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 14 (pool-1-thread-1): State: RUNNABLE Blocked count: 0 Waited count: 25 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.poll(EPollPort.java:200) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:281) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 15 (pool-1-thread-2): State: WAITING Blocked count: 0 Waited count: 23 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@31a5b7ac Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:275) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 16 (surefire-forkedjvm-stream-flusher): State: TIMED_WAITING Blocked count: 0 Waited count: 4578 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 18 (surefire-forkedjvm-command-thread): State: WAITING Blocked count: 0 Waited count: 47 Waiting on java.util.concurrent.CountDownLatch$Sync@17f550cc Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.CountDownLatch.await(CountDownLatch.java:230) java.base@17.0.11/sun.nio.ch.PendingFuture.get(PendingFuture.java:178) app//org.apache.maven.surefire.api.util.internal.Channels$2.read(Channels.java:127) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:284) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) app//org.apache.maven.surefire.api.util.internal.Channels$3.readImpl(Channels.java:169) app//org.apache.maven.surefire.api.util.internal.AbstractNoninterruptibleReadableChannel.read(AbstractNoninterruptibleReadableChannel.java:50) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:430) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:419) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.readMessageType(AbstractStreamDecoder.java:116) app//org.apache.maven.surefire.booter.stream.CommandDecoder.decode(CommandDecoder.java:77) app//org.apache.maven.surefire.booter.spi.CommandChannelDecoder.decode(CommandChannelDecoder.java:60) app//org.apache.maven.surefire.booter.CommandReader$CommandRunnable.run(CommandReader.java:290) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 22 (Time-limited test): State: RUNNABLE Blocked count: 12870 Waited count: 13419 Stack: java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo1(Native Method) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:197) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:154) app//org.apache.hadoop.hbase.util.ReflectionUtils.printThreadInfo(ReflectionUtils.java:181) app//org.apache.hadoop.hbase.util.Threads.printThreadInfo(Threads.java:186) app//org.apache.hadoop.hbase.util.Threads.threadDumpingIsAlive(Threads.java:113) app//org.apache.hadoop.hbase.LocalHBaseCluster.join(LocalHBaseCluster.java:394) app//org.apache.hadoop.hbase.MiniHBaseCluster.waitUntilShutDown(MiniHBaseCluster.java:921) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniHBaseCluster(HBaseTestingUtility.java:1359) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniCluster(HBaseTestingUtility.java:1341) app//org.apache.hadoop.hbase.snapshot.TestExportSnapshot.tearDownAfterClass(TestExportSnapshot.java:121) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) java.base@17.0.11/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) java.base@17.0.11/java.lang.reflect.Method.invoke(Method.java:568) app//org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) app//org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) app//org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) app//org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) app//org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) Thread 23 (org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner): State: WAITING Blocked count: 12 Waited count: 13 Waiting on java.lang.ref.ReferenceQueue$Lock@2fdea105 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) app//org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner.run(FileSystem.java:4171) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 25 (SSL Certificates Store Monitor): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.TaskQueue@33da7b85 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 34 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@44d16179): State: TIMED_WAITING Blocked count: 0 Waited count: 909 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 35 (GcTimeMonitor obsWindow = 60000, sleepInterval = 5000, maxGcTimePerc = 100): State: TIMED_WAITING Blocked count: 0 Waited count: 91 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.GcTimeMonitor.run(GcTimeMonitor.java:161) Thread 36 (pool-6-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 37 (qtp123845481-37): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 38 (qtp123845481-38): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 39 (qtp123845481-39): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 40 (qtp123845481-40): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 41 (qtp123845481-41-acceptor-0@781b1e42-ServerConnector@67ec1049{HTTP/1.1, (http/1.1)}{localhost:42647}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 42 (qtp123845481-42): State: TIMED_WAITING Blocked count: 0 Waited count: 8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 43 (qtp123845481-43): State: TIMED_WAITING Blocked count: 0 Waited count: 8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 44 (qtp123845481-44): State: TIMED_WAITING Blocked count: 0 Waited count: 8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 45 (Session-HouseKeeper-604db391-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 46 (pool-7-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 53 (FSEditLogAsync): State: WAITING Blocked count: 28 Waited count: 3355 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@60a45dc1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.dequeueEdit(FSEditLogAsync.java:241) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.run(FSEditLogAsync.java:250) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 55 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 56 (IPC Server idle connection scanner for port 37495): State: TIMED_WAITING Blocked count: 1 Waited count: 47 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 58 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 91 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 61 (org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor@33e58e69): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor.run(PendingReconstructionBlocks.java:267) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 62 (DatanodeAdminMonitor-0): State: TIMED_WAITING Blocked count: 0 Waited count: 152 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 49 (org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor@989c5e0): State: TIMED_WAITING Blocked count: 0 Waited count: 91 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor.run(HeartbeatManager.java:563) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 47 (RedundancyMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 156 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) java.base@17.0.11/java.lang.Thread.sleep(Thread.java:344) java.base@17.0.11/java.util.concurrent.TimeUnit.sleep(TimeUnit.java:446) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$RedundancyMonitor.run(BlockManager.java:5352) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 48 (MarkedDeleteBlockScrubberThread): State: TIMED_WAITING Blocked count: 0 Waited count: 44877 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$MarkedDeleteBlockScrubber.run(BlockManager.java:5326) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 51 (Block report processor): State: WAITING Blocked count: 1 Waited count: 1625 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@468106b7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.processQueue(BlockManager.java:5627) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.run(BlockManager.java:5614) Thread 57 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 54 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 64 (IPC Server handler 0 on default port 37495): State: TIMED_WAITING Blocked count: 70 Waited count: 2441 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 65 (IPC Server handler 1 on default port 37495): State: TIMED_WAITING Blocked count: 70 Waited count: 2440 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 66 (IPC Server handler 2 on default port 37495): State: TIMED_WAITING Blocked count: 66 Waited count: 2444 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 67 (IPC Server handler 3 on default port 37495): State: TIMED_WAITING Blocked count: 66 Waited count: 2440 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 68 (IPC Server handler 4 on default port 37495): State: TIMED_WAITING Blocked count: 82 Waited count: 2428 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 69 (pool-12-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 71 (org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor@7d0a4f0): State: TIMED_WAITING Blocked count: 0 Waited count: 227 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor.run(LeaseManager.java:537) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 72 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor@108b955): State: TIMED_WAITING Blocked count: 0 Waited count: 91 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor.run(FSNamesystem.java:4550) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 73 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller@2ea8d10a): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller.run(FSNamesystem.java:4592) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 74 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber@5985a9b3): State: TIMED_WAITING Blocked count: 0 Waited count: 4 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber.run(FSNamesystem.java:4689) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 75 (CacheReplicationMonitor(822228812)): State: TIMED_WAITING Blocked count: 0 Waited count: 17 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1759) app//org.apache.hadoop.hdfs.server.blockmanagement.CacheReplicationMonitor.run(CacheReplicationMonitor.java:186) Thread 85 (pool-18-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 86 (qtp1135609267-86): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 87 (qtp1135609267-87-acceptor-0@473ea113-ServerConnector@3c5cb06a{HTTP/1.1, (http/1.1)}{localhost:34891}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 88 (qtp1135609267-88): State: TIMED_WAITING Blocked count: 0 Waited count: 8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 89 (qtp1135609267-89): State: TIMED_WAITING Blocked count: 0 Waited count: 8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 90 (Session-HouseKeeper-4ee4d1fd-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 91 (nioEventLoopGroup-2-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 92 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@3900f107): State: TIMED_WAITING Blocked count: 0 Waited count: 906 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 94 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 95 (IPC Server idle connection scanner for port 45989): State: TIMED_WAITING Blocked count: 1 Waited count: 47 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 97 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 91 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 100 (Command processor): State: WAITING Blocked count: 3 Waited count: 296 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@18778abc Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 101 (BP-22138141-172.17.0.2-1736333101274 heartbeating to localhost/127.0.0.1:37495): State: TIMED_WAITING Blocked count: 1378 Waited count: 1536 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 102 (pool-20-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 84 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@8acd5ca): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 96 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 93 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 103 (IPC Server handler 0 on default port 45989): State: TIMED_WAITING Blocked count: 0 Waited count: 453 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 104 (IPC Server handler 1 on default port 45989): State: TIMED_WAITING Blocked count: 0 Waited count: 453 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 105 (IPC Server handler 2 on default port 45989): State: TIMED_WAITING Blocked count: 0 Waited count: 459 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 106 (IPC Server handler 3 on default port 45989): State: TIMED_WAITING Blocked count: 0 Waited count: 467 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 109 (IPC Server handler 4 on default port 45989): State: TIMED_WAITING Blocked count: 0 Waited count: 453 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 107 (IPC Client (175901922) connection to localhost/127.0.0.1:37495 from jenkins): State: TIMED_WAITING Blocked count: 1425 Waited count: 1426 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Thread 108 (IPC Parameter Sending Thread for localhost/127.0.0.1:37495): State: TIMED_WAITING Blocked count: 0 Waited count: 2180 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 119 (pool-26-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 120 (qtp40190603-120): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 121 (qtp40190603-121-acceptor-0@733970c9-ServerConnector@27b3056c{HTTP/1.1, (http/1.1)}{localhost:42931}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 122 (qtp40190603-122): State: TIMED_WAITING Blocked count: 0 Waited count: 8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 123 (qtp40190603-123): State: TIMED_WAITING Blocked count: 0 Waited count: 8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 124 (Session-HouseKeeper-66a6431-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 127 (nioEventLoopGroup-4-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 128 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@63d3adf1): State: TIMED_WAITING Blocked count: 0 Waited count: 904 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 130 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 131 (IPC Server idle connection scanner for port 37919): State: TIMED_WAITING Blocked count: 1 Waited count: 47 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 133 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 91 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 136 (Command processor): State: WAITING Blocked count: 0 Waited count: 277 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@262fd62 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 137 (BP-22138141-172.17.0.2-1736333101274 heartbeating to localhost/127.0.0.1:37495): State: TIMED_WAITING Blocked count: 1399 Waited count: 1538 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 138 (pool-29-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 118 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@3e496c77): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 132 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 129 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 139 (IPC Server handler 0 on default port 37919): State: TIMED_WAITING Blocked count: 0 Waited count: 452 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 140 (IPC Server handler 1 on default port 37919): State: TIMED_WAITING Blocked count: 0 Waited count: 452 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 141 (IPC Server handler 2 on default port 37919): State: TIMED_WAITING Blocked count: 0 Waited count: 452 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 142 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data1)): State: TIMED_WAITING Blocked count: 7 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 144 (IPC Server handler 3 on default port 37919): State: TIMED_WAITING Blocked count: 0 Waited count: 452 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 146 (IPC Server handler 4 on default port 37919): State: TIMED_WAITING Blocked count: 0 Waited count: 452 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 145 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data2)): State: TIMED_WAITING Blocked count: 10 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 163 (pool-40-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 158 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data1/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 159 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data3)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 164 (qtp1264850999-164): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 165 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data4)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 166 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data2/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 167 (qtp1264850999-167-acceptor-0@758b1918-ServerConnector@3362f87c{HTTP/1.1, (http/1.1)}{localhost:40601}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 171 (qtp1264850999-171): State: TIMED_WAITING Blocked count: 0 Waited count: 8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 172 (qtp1264850999-172): State: TIMED_WAITING Blocked count: 0 Waited count: 8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 176 (Session-HouseKeeper-26f1d292-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 180 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data3/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 181 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data4/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 183 (pool-15-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 186 (java.util.concurrent.ThreadPoolExecutor$Worker@2c953ef0[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 190 (pool-23-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 193 (java.util.concurrent.ThreadPoolExecutor$Worker@1c56948b[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 194 (nioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 195 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@6eba8751): State: TIMED_WAITING Blocked count: 0 Waited count: 903 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 197 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 198 (IPC Server idle connection scanner for port 45913): State: TIMED_WAITING Blocked count: 1 Waited count: 47 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 200 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 91 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 203 (Command processor): State: WAITING Blocked count: 1 Waited count: 315 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@336893f4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 204 (BP-22138141-172.17.0.2-1736333101274 heartbeating to localhost/127.0.0.1:37495): State: TIMED_WAITING Blocked count: 1365 Waited count: 1531 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 205 (pool-46-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 162 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@7b35f50f): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 199 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 196 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 206 (IPC Server handler 0 on default port 45913): State: TIMED_WAITING Blocked count: 0 Waited count: 452 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 207 (IPC Server handler 1 on default port 45913): State: TIMED_WAITING Blocked count: 0 Waited count: 452 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 208 (IPC Server handler 2 on default port 45913): State: TIMED_WAITING Blocked count: 0 Waited count: 452 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 209 (IPC Server handler 3 on default port 45913): State: TIMED_WAITING Blocked count: 0 Waited count: 453 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 210 (IPC Server handler 4 on default port 45913): State: TIMED_WAITING Blocked count: 0 Waited count: 453 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 213 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data5)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 214 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data6)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 220 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data5/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 221 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data6/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 225 (pool-37-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 228 (java.util.concurrent.ThreadPoolExecutor$Worker@5fd6a7f8[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 233 (FsDatasetAsyncDiskServiceFixer): State: TIMED_WAITING Blocked count: 0 Waited count: 16 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer.run(HBaseTestingUtility.java:599) Thread 236 (NIOServerCxnFactory.SelectorThread-1): State: RUNNABLE Blocked count: 2 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 235 (NIOServerCxnFactory.SelectorThread-0): State: RUNNABLE Blocked count: 16 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 237 (NIOServerCxnFactory.AcceptThread:localhost/127.0.0.1:52367): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.select(NIOServerCnxnFactory.java:205) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.run(NIOServerCnxnFactory.java:181) Thread 234 (ConnnectionExpirer): State: TIMED_WAITING Blocked count: 0 Waited count: 46 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.NIOServerCnxnFactory$ConnectionExpirerThread.run(NIOServerCnxnFactory.java:554) Thread 238 (SessionTracker): State: TIMED_WAITING Blocked count: 0 Waited count: 226 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.SessionTrackerImpl.run(SessionTrackerImpl.java:163) Thread 239 (SyncThread:0): State: WAITING Blocked count: 33 Waited count: 719 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4dc1adaf Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.SyncRequestProcessor.run(SyncRequestProcessor.java:170) Thread 240 (ProcessThread(sid:0 cport:52367):): State: WAITING Blocked count: 0 Waited count: 809 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@d2bfd92 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.PrepRequestProcessor.run(PrepRequestProcessor.java:142) Thread 241 (RequestThrottler): State: WAITING Blocked count: 0 Waited count: 842 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@3c783c72 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.RequestThrottler.run(RequestThrottler.java:147) Thread 242 (NIOWorkerThread-1): State: WAITING Blocked count: 0 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 253 (Time-limited test.named-queue-events-pool-0): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@ff82e77 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) app//com.lmax.disruptor.BlockingWaitStrategy.waitFor(BlockingWaitStrategy.java:47) app//com.lmax.disruptor.ProcessingSequenceBarrier.waitFor(ProcessingSequenceBarrier.java:56) app//com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:159) app//com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 254 (HBase-Metrics2-1): State: TIMED_WAITING Blocked count: 0 Waited count: 320 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 255 (RS-EventLoopGroup-1-1): State: RUNNABLE Blocked count: 50 Waited count: 1 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 256 (Time-limited test-SendThread(127.0.0.1:52367)): State: RUNNABLE Blocked count: 20 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:332) app//org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1289) Thread 257 (Time-limited test-EventThread): State: WAITING Blocked count: 2 Waited count: 58 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@200ba914 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:550) Thread 258 (NIOWorkerThread-2): State: WAITING Blocked count: 3 Waited count: 136 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 259 (NIOWorkerThread-3): State: WAITING Blocked count: 3 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 260 (NIOWorkerThread-4): State: WAITING Blocked count: 1 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 261 (NIOWorkerThread-5): State: WAITING Blocked count: 3 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 262 (zk-event-processor-pool-0): State: WAITING Blocked count: 35 Waited count: 87 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4d05f8cc Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 263 (NIOWorkerThread-6): State: WAITING Blocked count: 4 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 264 (NIOWorkerThread-7): State: WAITING Blocked count: 3 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 265 (NIOWorkerThread-8): State: WAITING Blocked count: 0 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 266 (NIOWorkerThread-9): State: WAITING Blocked count: 2 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 267 (NIOWorkerThread-10): State: WAITING Blocked count: 3 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 268 (NIOWorkerThread-11): State: WAITING Blocked count: 6 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 269 (NIOWorkerThread-12): State: WAITING Blocked count: 1 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 270 (NIOWorkerThread-13): State: WAITING Blocked count: 2 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 271 (NIOWorkerThread-14): State: WAITING Blocked count: 3 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 272 (NIOWorkerThread-15): State: WAITING Blocked count: 2 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 273 (NIOWorkerThread-16): State: WAITING Blocked count: 2 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 274 (RpcServer.default.FPBQ.Fifo.handler=0,queue=0,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@3c0ec8d Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 275 (RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551): State: WAITING Blocked count: 221 Waited count: 834 Waiting on java.util.concurrent.Semaphore$NonfairSync@32b787af Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 276 (RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551): State: WAITING Blocked count: 93 Waited count: 356 Waiting on java.util.concurrent.Semaphore$NonfairSync@4f89cbe4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 277 (RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=34551): State: WAITING Blocked count: 80 Waited count: 5313 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@2d289941 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 278 (RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@2988bced Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 279 (RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@2988bced Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 280 (RpcServer.replication.FPBQ.Fifo.handler=0,queue=0,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@278ccc8c Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 281 (RpcServer.replication.FPBQ.Fifo.handler=1,queue=0,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@265be572 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 282 (RpcServer.replication.FPBQ.Fifo.handler=2,queue=0,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@4fd08269 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 283 (RpcServer.metaPriority.FPBQ.Fifo.handler=0,queue=0,port=34551): State: WAITING Blocked count: 0 Waited count: 3 Waiting on java.util.concurrent.Semaphore$NonfairSync@23d320aa Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 287 (RS-EventLoopGroup-3-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 307 (RS-EventLoopGroup-4-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 327 (RS-EventLoopGroup-5-1): State: RUNNABLE Blocked count: 90 Waited count: 4 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 285 (M:0;67f42ee69b4e:34551): State: TIMED_WAITING Blocked count: 6 Waited count: 2682 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1759) app//org.apache.hadoop.hbase.regionserver.wal.SyncFuture.get(SyncFuture.java:169) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.blockOnSync(AbstractFSWAL.java:879) app//org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.doSync(AsyncFSWAL.java:671) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.lambda$sync$1(AbstractFSWAL.java:600) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL$$Lambda$983/0x00007f478cf19800.run(Unknown Source) app//org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:600) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:590) app//org.apache.hadoop.hbase.regionserver.HRegion.doSyncOfUnflushedWALChanges(HRegion.java:2869) app//org.apache.hadoop.hbase.regionserver.HRegion.internalPrepareFlushCache(HRegion.java:2811) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2670) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2644) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2635) app//org.apache.hadoop.hbase.regionserver.HRegion.doClose(HRegion.java:1810) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1631) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1586) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1569) Thread 348 (Monitor thread for TaskMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 45 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.monitoring.TaskMonitor$MonitorRunnable.run(TaskMonitor.java:325) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 350 (master/67f42ee69b4e:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 352 (master/67f42ee69b4e:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 354 (org.apache.hadoop.hdfs.PeerCache@58490a7b): State: TIMED_WAITING Blocked count: 0 Waited count: 150 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.PeerCache.run(PeerCache.java:253) app//org.apache.hadoop.hdfs.PeerCache.access$000(PeerCache.java:46) app//org.apache.hadoop.hdfs.PeerCache$1.run(PeerCache.java:124) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 373 (master:store-WAL-Roller): State: TIMED_WAITING Blocked count: 0 Waited count: 4473 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hbase.wal.AbstractWALRoller.run(AbstractWALRoller.java:179) Thread 390 (RS-EventLoopGroup-5-2): State: RUNNABLE Blocked count: 108 Waited count: 3 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 391 (RS-EventLoopGroup-5-3): State: RUNNABLE Blocked count: 131 Waited count: 3 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 404 (Idle-Rpc-Conn-Sweeper-pool-0): State: WAITING Blocked count: 0 Waited count: 65 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6a97dfa0 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 415 (SnapshotHandlerChoreCleaner): State: TIMED_WAITING Blocked count: 0 Waited count: 45 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 403 (RpcClient-timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 44684 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 424 (RS-EventLoopGroup-1-2): State: RUNNABLE Blocked count: 27 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 425 (RS-EventLoopGroup-1-3): State: RUNNABLE Blocked count: 56 Waited count: 2 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 450 (RegionServerTracker-0): State: WAITING Blocked count: 8 Waited count: 10 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5be54308 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 472 (regionserver/67f42ee69b4e:0.procedureResultReporter): State: WAITING Blocked count: 14 Waited count: 29 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@508e73da Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 473 (regionserver/67f42ee69b4e:0.procedureResultReporter): State: WAITING Blocked count: 22 Waited count: 33 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@430dcb4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 474 (regionserver/67f42ee69b4e:0.procedureResultReporter): State: WAITING Blocked count: 14 Waited count: 23 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@595b70f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 512 (RS-EventLoopGroup-4-2): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 523 (RS-EventLoopGroup-4-3): State: RUNNABLE Blocked count: 2 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 525 (region-location-0): State: WAITING Blocked count: 8 Waited count: 12 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@741949b6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 551 (Async-Client-Retry-Timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 44457 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 566 (RPCClient-NioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 3 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 606 (region-location-1): State: WAITING Blocked count: 3 Waited count: 7 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@741949b6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 607 (region-location-2): State: WAITING Blocked count: 3 Waited count: 6 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@741949b6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 608 (region-location-3): State: WAITING Blocked count: 2 Waited count: 5 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@741949b6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1012 (MutableQuantiles-0): State: TIMED_WAITING Blocked count: 0 Waited count: 371 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1073 (RPCClient-NioEventLoopGroup-6-2): State: RUNNABLE Blocked count: 6 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1101 (RS-EventLoopGroup-3-2): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1112 (zk-permission-watcher-pool-0): State: WAITING Blocked count: 62 Waited count: 97 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7d78af0e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1120 (RPCClient-NioEventLoopGroup-6-3): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1121 (RS-EventLoopGroup-3-3): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1562 (Container metrics unregistration): State: WAITING Blocked count: 10 Waited count: 33 Waiting on java.util.TaskQueue@43921dee Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 2944 (RPCClient-NioEventLoopGroup-6-4): State: RUNNABLE Blocked count: 2 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 3542 (region-location-4): State: WAITING Blocked count: 3 Waited count: 6 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@741949b6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5203 (RPCClient-NioEventLoopGroup-6-5): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5204 (RPCClient-NioEventLoopGroup-6-6): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5205 (RPCClient-NioEventLoopGroup-6-7): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9395 (AsyncFSWAL-1-hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/MasterData-prefix:67f42ee69b4e,34551,1736333106957): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@431a8944 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9399 (Timer for 'JobHistoryServer' metrics system): State: TIMED_WAITING Blocked count: 0 Waited count: 15 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) 2025-01-08T10:53:05,716 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T10:53:35,717 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. Process Thread Dump: Automatic Stack Trace every 60 seconds waiting on M:0;67f42ee69b4e:34551 218 active threads Thread 1 (main): State: TIMED_WAITING Blocked count: 2 Waited count: 4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.FutureTask.awaitDone(FutureTask.java:444) java.base@17.0.11/java.util.concurrent.FutureTask.get(FutureTask.java:203) app//org.junit.internal.runners.statements.FailOnTimeout.getResult(FailOnTimeout.java:167) app//org.junit.internal.runners.statements.FailOnTimeout.evaluate(FailOnTimeout.java:128) app//org.apache.hadoop.hbase.SystemExitRule$1.evaluate(SystemExitRule.java:39) app//org.junit.rules.RunRules.evaluate(RunRules.java:20) app//org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) app//org.junit.runners.ParentRunner.run(ParentRunner.java:413) app//org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:316) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:240) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:214) app//org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:155) app//org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:385) app//org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) app//org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:507) app//org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:495) Thread 2 (Reference Handler): State: RUNNABLE Blocked count: 3 Waited count: 0 Stack: java.base@17.0.11/java.lang.ref.Reference.waitForReferencePendingList(Native Method) java.base@17.0.11/java.lang.ref.Reference.processPendingReferences(Reference.java:253) java.base@17.0.11/java.lang.ref.Reference$ReferenceHandler.run(Reference.java:215) Thread 3 (Finalizer): State: WAITING Blocked count: 40 Waited count: 12 Waiting on java.lang.ref.ReferenceQueue$Lock@199c2830 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) java.base@17.0.11/java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:172) Thread 4 (Signal Dispatcher): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 12 (Common-Cleaner): State: TIMED_WAITING Blocked count: 18 Waited count: 19 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/jdk.internal.ref.CleanerImpl.run(CleanerImpl.java:140) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) java.base@17.0.11/jdk.internal.misc.InnocuousThread.run(InnocuousThread.java:162) Thread 13 (Notification Thread): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 14 (pool-1-thread-1): State: WAITING Blocked count: 0 Waited count: 29 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@31a5b7ac Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:275) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 15 (pool-1-thread-2): State: RUNNABLE Blocked count: 0 Waited count: 25 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.poll(EPollPort.java:200) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:281) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 16 (surefire-forkedjvm-stream-flusher): State: TIMED_WAITING Blocked count: 0 Waited count: 5177 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 18 (surefire-forkedjvm-command-thread): State: WAITING Blocked count: 0 Waited count: 53 Waiting on java.util.concurrent.CountDownLatch$Sync@b2bbdf Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.CountDownLatch.await(CountDownLatch.java:230) java.base@17.0.11/sun.nio.ch.PendingFuture.get(PendingFuture.java:178) app//org.apache.maven.surefire.api.util.internal.Channels$2.read(Channels.java:127) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:284) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) app//org.apache.maven.surefire.api.util.internal.Channels$3.readImpl(Channels.java:169) app//org.apache.maven.surefire.api.util.internal.AbstractNoninterruptibleReadableChannel.read(AbstractNoninterruptibleReadableChannel.java:50) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:430) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:419) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.readMessageType(AbstractStreamDecoder.java:116) app//org.apache.maven.surefire.booter.stream.CommandDecoder.decode(CommandDecoder.java:77) app//org.apache.maven.surefire.booter.spi.CommandChannelDecoder.decode(CommandChannelDecoder.java:60) app//org.apache.maven.surefire.booter.CommandReader$CommandRunnable.run(CommandReader.java:290) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 22 (Time-limited test): State: RUNNABLE Blocked count: 12870 Waited count: 13420 Stack: java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo1(Native Method) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:197) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:154) app//org.apache.hadoop.hbase.util.ReflectionUtils.printThreadInfo(ReflectionUtils.java:181) app//org.apache.hadoop.hbase.util.Threads.printThreadInfo(Threads.java:186) app//org.apache.hadoop.hbase.util.Threads.threadDumpingIsAlive(Threads.java:113) app//org.apache.hadoop.hbase.LocalHBaseCluster.join(LocalHBaseCluster.java:394) app//org.apache.hadoop.hbase.MiniHBaseCluster.waitUntilShutDown(MiniHBaseCluster.java:921) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniHBaseCluster(HBaseTestingUtility.java:1359) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniCluster(HBaseTestingUtility.java:1341) app//org.apache.hadoop.hbase.snapshot.TestExportSnapshot.tearDownAfterClass(TestExportSnapshot.java:121) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) java.base@17.0.11/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) java.base@17.0.11/java.lang.reflect.Method.invoke(Method.java:568) app//org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) app//org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) app//org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) app//org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) app//org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) Thread 23 (org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner): State: WAITING Blocked count: 12 Waited count: 13 Waiting on java.lang.ref.ReferenceQueue$Lock@2fdea105 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) app//org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner.run(FileSystem.java:4171) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 25 (SSL Certificates Store Monitor): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.TaskQueue@33da7b85 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 34 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@44d16179): State: TIMED_WAITING Blocked count: 0 Waited count: 1029 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 35 (GcTimeMonitor obsWindow = 60000, sleepInterval = 5000, maxGcTimePerc = 100): State: TIMED_WAITING Blocked count: 0 Waited count: 103 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.GcTimeMonitor.run(GcTimeMonitor.java:161) Thread 36 (pool-6-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 37 (qtp123845481-37): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 38 (qtp123845481-38): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 39 (qtp123845481-39): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 40 (qtp123845481-40): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 41 (qtp123845481-41-acceptor-0@781b1e42-ServerConnector@67ec1049{HTTP/1.1, (http/1.1)}{localhost:42647}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 42 (qtp123845481-42): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 43 (qtp123845481-43): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 44 (qtp123845481-44): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 45 (Session-HouseKeeper-604db391-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 46 (pool-7-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 53 (FSEditLogAsync): State: WAITING Blocked count: 28 Waited count: 3355 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@60a45dc1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.dequeueEdit(FSEditLogAsync.java:241) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.run(FSEditLogAsync.java:250) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 55 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 56 (IPC Server idle connection scanner for port 37495): State: TIMED_WAITING Blocked count: 1 Waited count: 53 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 58 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 103 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 61 (org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor@33e58e69): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor.run(PendingReconstructionBlocks.java:267) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 62 (DatanodeAdminMonitor-0): State: TIMED_WAITING Blocked count: 0 Waited count: 172 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 49 (org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor@989c5e0): State: TIMED_WAITING Blocked count: 0 Waited count: 103 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor.run(HeartbeatManager.java:563) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 47 (RedundancyMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 176 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) java.base@17.0.11/java.lang.Thread.sleep(Thread.java:344) java.base@17.0.11/java.util.concurrent.TimeUnit.sleep(TimeUnit.java:446) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$RedundancyMonitor.run(BlockManager.java:5352) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 48 (MarkedDeleteBlockScrubberThread): State: TIMED_WAITING Blocked count: 0 Waited count: 50828 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$MarkedDeleteBlockScrubber.run(BlockManager.java:5326) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 51 (Block report processor): State: WAITING Blocked count: 1 Waited count: 1625 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@468106b7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.processQueue(BlockManager.java:5627) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.run(BlockManager.java:5614) Thread 57 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 54 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 64 (IPC Server handler 0 on default port 37495): State: TIMED_WAITING Blocked count: 70 Waited count: 2502 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 65 (IPC Server handler 1 on default port 37495): State: TIMED_WAITING Blocked count: 70 Waited count: 2502 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 66 (IPC Server handler 2 on default port 37495): State: TIMED_WAITING Blocked count: 66 Waited count: 2505 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 67 (IPC Server handler 3 on default port 37495): State: TIMED_WAITING Blocked count: 66 Waited count: 2500 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 68 (IPC Server handler 4 on default port 37495): State: TIMED_WAITING Blocked count: 82 Waited count: 2489 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 69 (pool-12-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 71 (org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor@7d0a4f0): State: TIMED_WAITING Blocked count: 0 Waited count: 257 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor.run(LeaseManager.java:537) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 72 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor@108b955): State: TIMED_WAITING Blocked count: 0 Waited count: 103 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor.run(FSNamesystem.java:4550) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 73 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller@2ea8d10a): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller.run(FSNamesystem.java:4592) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 74 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber@5985a9b3): State: TIMED_WAITING Blocked count: 0 Waited count: 4 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber.run(FSNamesystem.java:4689) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 75 (CacheReplicationMonitor(822228812)): State: TIMED_WAITING Blocked count: 0 Waited count: 19 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1759) app//org.apache.hadoop.hdfs.server.blockmanagement.CacheReplicationMonitor.run(CacheReplicationMonitor.java:186) Thread 85 (pool-18-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 86 (qtp1135609267-86): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 87 (qtp1135609267-87-acceptor-0@473ea113-ServerConnector@3c5cb06a{HTTP/1.1, (http/1.1)}{localhost:34891}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 88 (qtp1135609267-88): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 89 (qtp1135609267-89): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 90 (Session-HouseKeeper-4ee4d1fd-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 91 (nioEventLoopGroup-2-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 92 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@3900f107): State: TIMED_WAITING Blocked count: 0 Waited count: 1025 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 94 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 95 (IPC Server idle connection scanner for port 45989): State: TIMED_WAITING Blocked count: 1 Waited count: 53 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 97 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 103 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 100 (Command processor): State: WAITING Blocked count: 3 Waited count: 316 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@18778abc Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 101 (BP-22138141-172.17.0.2-1736333101274 heartbeating to localhost/127.0.0.1:37495): State: TIMED_WAITING Blocked count: 1398 Waited count: 1576 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 102 (pool-20-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 84 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@8acd5ca): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 96 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 93 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 103 (IPC Server handler 0 on default port 45989): State: TIMED_WAITING Blocked count: 0 Waited count: 513 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 104 (IPC Server handler 1 on default port 45989): State: TIMED_WAITING Blocked count: 0 Waited count: 513 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 105 (IPC Server handler 2 on default port 45989): State: TIMED_WAITING Blocked count: 0 Waited count: 520 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 106 (IPC Server handler 3 on default port 45989): State: TIMED_WAITING Blocked count: 0 Waited count: 528 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 109 (IPC Server handler 4 on default port 45989): State: TIMED_WAITING Blocked count: 0 Waited count: 513 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 107 (IPC Client (175901922) connection to localhost/127.0.0.1:37495 from jenkins): State: TIMED_WAITING Blocked count: 1485 Waited count: 1486 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Thread 108 (IPC Parameter Sending Thread for localhost/127.0.0.1:37495): State: TIMED_WAITING Blocked count: 0 Waited count: 2240 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 119 (pool-26-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 120 (qtp40190603-120): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 121 (qtp40190603-121-acceptor-0@733970c9-ServerConnector@27b3056c{HTTP/1.1, (http/1.1)}{localhost:42931}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 122 (qtp40190603-122): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 123 (qtp40190603-123): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 124 (Session-HouseKeeper-66a6431-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 127 (nioEventLoopGroup-4-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 128 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@63d3adf1): State: TIMED_WAITING Blocked count: 0 Waited count: 1024 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 130 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 131 (IPC Server idle connection scanner for port 37919): State: TIMED_WAITING Blocked count: 1 Waited count: 53 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 133 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 103 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 136 (Command processor): State: WAITING Blocked count: 0 Waited count: 297 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@262fd62 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 137 (BP-22138141-172.17.0.2-1736333101274 heartbeating to localhost/127.0.0.1:37495): State: TIMED_WAITING Blocked count: 1419 Waited count: 1578 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 138 (pool-29-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 118 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@3e496c77): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 132 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 129 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 139 (IPC Server handler 0 on default port 37919): State: TIMED_WAITING Blocked count: 0 Waited count: 512 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 140 (IPC Server handler 1 on default port 37919): State: TIMED_WAITING Blocked count: 0 Waited count: 512 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 141 (IPC Server handler 2 on default port 37919): State: TIMED_WAITING Blocked count: 0 Waited count: 512 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 142 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data1)): State: TIMED_WAITING Blocked count: 7 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 144 (IPC Server handler 3 on default port 37919): State: TIMED_WAITING Blocked count: 0 Waited count: 512 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 146 (IPC Server handler 4 on default port 37919): State: TIMED_WAITING Blocked count: 0 Waited count: 512 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 145 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data2)): State: TIMED_WAITING Blocked count: 10 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 163 (pool-40-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 158 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data1/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 159 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data3)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 164 (qtp1264850999-164): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 165 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data4)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 166 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data2/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 167 (qtp1264850999-167-acceptor-0@758b1918-ServerConnector@3362f87c{HTTP/1.1, (http/1.1)}{localhost:40601}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 171 (qtp1264850999-171): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 172 (qtp1264850999-172): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 176 (Session-HouseKeeper-26f1d292-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 180 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data3/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 181 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data4/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 183 (pool-15-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 186 (java.util.concurrent.ThreadPoolExecutor$Worker@2c953ef0[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 190 (pool-23-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 193 (java.util.concurrent.ThreadPoolExecutor$Worker@1c56948b[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 194 (nioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 195 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@6eba8751): State: TIMED_WAITING Blocked count: 0 Waited count: 1023 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 197 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 198 (IPC Server idle connection scanner for port 45913): State: TIMED_WAITING Blocked count: 1 Waited count: 53 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 200 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 103 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 203 (Command processor): State: WAITING Blocked count: 1 Waited count: 335 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@336893f4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 204 (BP-22138141-172.17.0.2-1736333101274 heartbeating to localhost/127.0.0.1:37495): State: TIMED_WAITING Blocked count: 1385 Waited count: 1571 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 205 (pool-46-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 162 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@7b35f50f): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 199 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 196 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 206 (IPC Server handler 0 on default port 45913): State: TIMED_WAITING Blocked count: 0 Waited count: 512 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 207 (IPC Server handler 1 on default port 45913): State: TIMED_WAITING Blocked count: 0 Waited count: 512 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 208 (IPC Server handler 2 on default port 45913): State: TIMED_WAITING Blocked count: 0 Waited count: 512 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 209 (IPC Server handler 3 on default port 45913): State: TIMED_WAITING Blocked count: 0 Waited count: 513 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 210 (IPC Server handler 4 on default port 45913): State: TIMED_WAITING Blocked count: 0 Waited count: 515 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 213 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data5)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 214 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data6)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 220 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data5/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 221 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data6/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 225 (pool-37-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 228 (java.util.concurrent.ThreadPoolExecutor$Worker@5fd6a7f8[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 233 (FsDatasetAsyncDiskServiceFixer): State: TIMED_WAITING Blocked count: 0 Waited count: 18 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer.run(HBaseTestingUtility.java:599) Thread 236 (NIOServerCxnFactory.SelectorThread-1): State: RUNNABLE Blocked count: 2 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 235 (NIOServerCxnFactory.SelectorThread-0): State: RUNNABLE Blocked count: 16 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 237 (NIOServerCxnFactory.AcceptThread:localhost/127.0.0.1:52367): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.select(NIOServerCnxnFactory.java:205) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.run(NIOServerCnxnFactory.java:181) Thread 234 (ConnnectionExpirer): State: TIMED_WAITING Blocked count: 0 Waited count: 52 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.NIOServerCnxnFactory$ConnectionExpirerThread.run(NIOServerCnxnFactory.java:554) Thread 238 (SessionTracker): State: TIMED_WAITING Blocked count: 0 Waited count: 256 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.SessionTrackerImpl.run(SessionTrackerImpl.java:163) Thread 239 (SyncThread:0): State: WAITING Blocked count: 33 Waited count: 724 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4dc1adaf Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.SyncRequestProcessor.run(SyncRequestProcessor.java:170) Thread 240 (ProcessThread(sid:0 cport:52367):): State: WAITING Blocked count: 0 Waited count: 814 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@d2bfd92 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.PrepRequestProcessor.run(PrepRequestProcessor.java:142) Thread 241 (RequestThrottler): State: WAITING Blocked count: 0 Waited count: 847 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@3c783c72 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.RequestThrottler.run(RequestThrottler.java:147) Thread 242 (NIOWorkerThread-1): State: WAITING Blocked count: 0 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 253 (Time-limited test.named-queue-events-pool-0): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@ff82e77 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) app//com.lmax.disruptor.BlockingWaitStrategy.waitFor(BlockingWaitStrategy.java:47) app//com.lmax.disruptor.ProcessingSequenceBarrier.waitFor(ProcessingSequenceBarrier.java:56) app//com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:159) app//com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 254 (HBase-Metrics2-1): State: TIMED_WAITING Blocked count: 0 Waited count: 348 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 255 (RS-EventLoopGroup-1-1): State: RUNNABLE Blocked count: 50 Waited count: 1 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 256 (Time-limited test-SendThread(127.0.0.1:52367)): State: RUNNABLE Blocked count: 20 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:332) app//org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1289) Thread 257 (Time-limited test-EventThread): State: WAITING Blocked count: 2 Waited count: 58 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@200ba914 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:550) Thread 258 (NIOWorkerThread-2): State: WAITING Blocked count: 3 Waited count: 137 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 259 (NIOWorkerThread-3): State: WAITING Blocked count: 3 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 260 (NIOWorkerThread-4): State: WAITING Blocked count: 1 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 261 (NIOWorkerThread-5): State: WAITING Blocked count: 3 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 262 (zk-event-processor-pool-0): State: WAITING Blocked count: 35 Waited count: 87 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4d05f8cc Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 263 (NIOWorkerThread-6): State: WAITING Blocked count: 4 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 264 (NIOWorkerThread-7): State: WAITING Blocked count: 3 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 265 (NIOWorkerThread-8): State: WAITING Blocked count: 0 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 266 (NIOWorkerThread-9): State: WAITING Blocked count: 2 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 267 (NIOWorkerThread-10): State: WAITING Blocked count: 3 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 268 (NIOWorkerThread-11): State: WAITING Blocked count: 6 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 269 (NIOWorkerThread-12): State: WAITING Blocked count: 1 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 270 (NIOWorkerThread-13): State: WAITING Blocked count: 2 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 271 (NIOWorkerThread-14): State: WAITING Blocked count: 3 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 272 (NIOWorkerThread-15): State: WAITING Blocked count: 2 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 273 (NIOWorkerThread-16): State: WAITING Blocked count: 2 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 274 (RpcServer.default.FPBQ.Fifo.handler=0,queue=0,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@3c0ec8d Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 275 (RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551): State: WAITING Blocked count: 221 Waited count: 834 Waiting on java.util.concurrent.Semaphore$NonfairSync@32b787af Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 276 (RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551): State: WAITING Blocked count: 93 Waited count: 356 Waiting on java.util.concurrent.Semaphore$NonfairSync@4f89cbe4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 277 (RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=34551): State: WAITING Blocked count: 80 Waited count: 5313 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@2d289941 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 278 (RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@2988bced Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 279 (RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@2988bced Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 280 (RpcServer.replication.FPBQ.Fifo.handler=0,queue=0,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@278ccc8c Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 281 (RpcServer.replication.FPBQ.Fifo.handler=1,queue=0,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@265be572 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 282 (RpcServer.replication.FPBQ.Fifo.handler=2,queue=0,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@4fd08269 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 283 (RpcServer.metaPriority.FPBQ.Fifo.handler=0,queue=0,port=34551): State: WAITING Blocked count: 0 Waited count: 3 Waiting on java.util.concurrent.Semaphore$NonfairSync@23d320aa Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 287 (RS-EventLoopGroup-3-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 307 (RS-EventLoopGroup-4-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 327 (RS-EventLoopGroup-5-1): State: RUNNABLE Blocked count: 90 Waited count: 4 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 285 (M:0;67f42ee69b4e:34551): State: TIMED_WAITING Blocked count: 6 Waited count: 2682 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1759) app//org.apache.hadoop.hbase.regionserver.wal.SyncFuture.get(SyncFuture.java:169) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.blockOnSync(AbstractFSWAL.java:879) app//org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.doSync(AsyncFSWAL.java:671) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.lambda$sync$1(AbstractFSWAL.java:600) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL$$Lambda$983/0x00007f478cf19800.run(Unknown Source) app//org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:600) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:590) app//org.apache.hadoop.hbase.regionserver.HRegion.doSyncOfUnflushedWALChanges(HRegion.java:2869) app//org.apache.hadoop.hbase.regionserver.HRegion.internalPrepareFlushCache(HRegion.java:2811) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2670) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2644) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2635) app//org.apache.hadoop.hbase.regionserver.HRegion.doClose(HRegion.java:1810) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1631) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1586) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1569) Thread 348 (Monitor thread for TaskMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 51 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.monitoring.TaskMonitor$MonitorRunnable.run(TaskMonitor.java:325) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 350 (master/67f42ee69b4e:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 352 (master/67f42ee69b4e:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 354 (org.apache.hadoop.hdfs.PeerCache@58490a7b): State: TIMED_WAITING Blocked count: 0 Waited count: 170 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.PeerCache.run(PeerCache.java:253) app//org.apache.hadoop.hdfs.PeerCache.access$000(PeerCache.java:46) app//org.apache.hadoop.hdfs.PeerCache$1.run(PeerCache.java:124) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 373 (master:store-WAL-Roller): State: TIMED_WAITING Blocked count: 0 Waited count: 5072 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hbase.wal.AbstractWALRoller.run(AbstractWALRoller.java:179) Thread 390 (RS-EventLoopGroup-5-2): State: RUNNABLE Blocked count: 108 Waited count: 3 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 391 (RS-EventLoopGroup-5-3): State: RUNNABLE Blocked count: 131 Waited count: 3 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 404 (Idle-Rpc-Conn-Sweeper-pool-0): State: WAITING Blocked count: 0 Waited count: 65 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6a97dfa0 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 415 (SnapshotHandlerChoreCleaner): State: TIMED_WAITING Blocked count: 0 Waited count: 51 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 403 (RpcClient-timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 50685 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 424 (RS-EventLoopGroup-1-2): State: RUNNABLE Blocked count: 27 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 425 (RS-EventLoopGroup-1-3): State: RUNNABLE Blocked count: 56 Waited count: 2 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 450 (RegionServerTracker-0): State: WAITING Blocked count: 8 Waited count: 10 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5be54308 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 472 (regionserver/67f42ee69b4e:0.procedureResultReporter): State: WAITING Blocked count: 14 Waited count: 29 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@508e73da Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 473 (regionserver/67f42ee69b4e:0.procedureResultReporter): State: WAITING Blocked count: 22 Waited count: 33 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@430dcb4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 474 (regionserver/67f42ee69b4e:0.procedureResultReporter): State: WAITING Blocked count: 14 Waited count: 23 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@595b70f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 512 (RS-EventLoopGroup-4-2): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 523 (RS-EventLoopGroup-4-3): State: RUNNABLE Blocked count: 2 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 525 (region-location-0): State: WAITING Blocked count: 8 Waited count: 12 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@741949b6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 551 (Async-Client-Retry-Timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 50459 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 566 (RPCClient-NioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 3 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 606 (region-location-1): State: WAITING Blocked count: 3 Waited count: 7 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@741949b6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 607 (region-location-2): State: WAITING Blocked count: 3 Waited count: 6 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@741949b6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 608 (region-location-3): State: WAITING Blocked count: 2 Waited count: 5 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@741949b6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1012 (MutableQuantiles-0): State: TIMED_WAITING Blocked count: 0 Waited count: 377 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1073 (RPCClient-NioEventLoopGroup-6-2): State: RUNNABLE Blocked count: 6 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1101 (RS-EventLoopGroup-3-2): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1112 (zk-permission-watcher-pool-0): State: WAITING Blocked count: 62 Waited count: 97 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7d78af0e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1120 (RPCClient-NioEventLoopGroup-6-3): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1121 (RS-EventLoopGroup-3-3): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1562 (Container metrics unregistration): State: WAITING Blocked count: 10 Waited count: 33 Waiting on java.util.TaskQueue@43921dee Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 2944 (RPCClient-NioEventLoopGroup-6-4): State: RUNNABLE Blocked count: 2 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 3542 (region-location-4): State: WAITING Blocked count: 3 Waited count: 6 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@741949b6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5203 (RPCClient-NioEventLoopGroup-6-5): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5204 (RPCClient-NioEventLoopGroup-6-6): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5205 (RPCClient-NioEventLoopGroup-6-7): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9395 (AsyncFSWAL-1-hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/MasterData-prefix:67f42ee69b4e,34551,1736333106957): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@431a8944 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9399 (Timer for 'JobHistoryServer' metrics system): State: TIMED_WAITING Blocked count: 0 Waited count: 21 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) 2025-01-08T10:54:05,717 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T10:54:35,718 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T10:54:36,438 DEBUG [M:0;67f42ee69b4e:34551 {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2025-01-08T10:54:36,438 WARN [M:0;67f42ee69b4e:34551 {}] region.MasterRegion(134): Failed to close region org.apache.hadoop.hbase.regionserver.wal.WALSyncTimeoutIOException: org.apache.hadoop.hbase.exceptions.TimeoutIOException: Failed to get sync result after 300000 ms for txid=3721, WAL system stuck? at org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.blockOnSync(AbstractFSWAL.java:883) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.doSync(AsyncFSWAL.java:671) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.lambda$sync$1(AbstractFSWAL.java:600) ~[classes/:?] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:600) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:590) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.doSyncOfUnflushedWALChanges(HRegion.java:2869) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.internalPrepareFlushCache(HRegion.java:2811) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2670) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2644) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2635) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.doClose(HRegion.java:1810) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1631) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1586) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1569) ~[classes/:?] at org.apache.hadoop.hbase.master.region.MasterRegion.closeRegion(MasterRegion.java:132) ~[classes/:?] at org.apache.hadoop.hbase.master.region.MasterRegion.close(MasterRegion.java:205) ~[classes/:?] at org.apache.hadoop.hbase.master.HMaster.stopServiceThreads(HMaster.java:1769) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:1285) ~[classes/:?] at org.apache.hadoop.hbase.master.HMaster.run(HMaster.java:614) ~[classes/:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: org.apache.hadoop.hbase.exceptions.TimeoutIOException: Failed to get sync result after 300000 ms for txid=3721, WAL system stuck? at org.apache.hadoop.hbase.regionserver.wal.SyncFuture.get(SyncFuture.java:171) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.blockOnSync(AbstractFSWAL.java:879) ~[classes/:?] ... 20 more 2025-01-08T10:54:36,440 WARN [Close-WAL-Writer-0 {}] wal.AsyncProtobufLogWriter(163): normal close failed, try recover java.lang.NullPointerException: Cannot invoke "org.apache.hbase.thirdparty.io.netty.buffer.ByteBuf.ensureWritable(int)" because "this.buf" is null at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutput.writeInt(FanOutOneBlockAsyncDFSOutput.java:396) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.wal.AsyncProtobufLogWriter.lambda$writeWALTrailerAndMagic$3(AsyncProtobufLogWriter.java:243) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncProtobufLogWriter.writeWALMetadata(AsyncProtobufLogWriter.java:201) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncProtobufLogWriter.writeWALTrailerAndMagic(AsyncProtobufLogWriter.java:236) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AbstractProtobufLogWriter.writeWALTrailer(AbstractProtobufLogWriter.java:252) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncProtobufLogWriter.close(AsyncProtobufLogWriter.java:160) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.lambda$closeWriter$5(AsyncFSWAL.java:732) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2025-01-08T10:54:36,444 DEBUG [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(47): Initialize RecoverLeaseFSUtils 2025-01-08T10:54:36,444 DEBUG [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(59): set recoverLeaseMethod to org.apache.hadoop.fs.LeaseRecoverable.recoverLease() 2025-01-08T10:54:36,444 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file /user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/MasterData/WALs/67f42ee69b4e,34551,1736333106957/67f42ee69b4e%2C34551%2C1736333106957.1736333108999 2025-01-08T10:54:36,444 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(233): attempt=0 on file=/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/MasterData/WALs/67f42ee69b4e,34551,1736333106957/67f42ee69b4e%2C34551%2C1736333106957.1736333108999 after 0ms java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.recoverLease(DFSClient.java:946) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:317) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:314) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.recoverLease(DistributedFileSystem.java:329) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverLease(RecoverLeaseFSUtils.java:222) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:155) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutput.recoverAndClose(FanOutOneBlockAsyncDFSOutput.java:610) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.wal.AsyncProtobufLogWriter.close(AsyncProtobufLogWriter.java:164) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.lambda$closeWriter$5(AsyncFSWAL.java:732) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2025-01-08T10:54:36,445 WARN [Close-WAL-Writer-0 {}] wal.AsyncFSWAL(734): close old writer failed. java.io.InterruptedIOException: Operation cancelled at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.checkIfCancelled(RecoverLeaseFSUtils.java:269) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:159) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutput.recoverAndClose(FanOutOneBlockAsyncDFSOutput.java:610) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.wal.AsyncProtobufLogWriter.close(AsyncProtobufLogWriter.java:164) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.lambda$closeWriter$5(AsyncFSWAL.java:732) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2025-01-08T10:54:36,445 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/MasterData/WALs/67f42ee69b4e,34551,1736333106957/67f42ee69b4e%2C34551%2C1736333106957.1736333108999 2025-01-08T10:54:36,445 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(233): attempt=0 on file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/MasterData/WALs/67f42ee69b4e,34551,1736333106957/67f42ee69b4e%2C34551%2C1736333106957.1736333108999 after 0ms java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.recoverLease(DFSClient.java:946) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:317) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:314) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.recoverLease(DistributedFileSystem.java:329) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverLease(RecoverLeaseFSUtils.java:222) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:155) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.recoverLease(AsyncFSWAL.java:722) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.lambda$closeWriter$5(AsyncFSWAL.java:735) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Process Thread Dump: Automatic Stack Trace every 60 seconds waiting on M:0;67f42ee69b4e:34551 221 active threads Thread 1 (main): State: TIMED_WAITING Blocked count: 2 Waited count: 4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.FutureTask.awaitDone(FutureTask.java:444) java.base@17.0.11/java.util.concurrent.FutureTask.get(FutureTask.java:203) app//org.junit.internal.runners.statements.FailOnTimeout.getResult(FailOnTimeout.java:167) app//org.junit.internal.runners.statements.FailOnTimeout.evaluate(FailOnTimeout.java:128) app//org.apache.hadoop.hbase.SystemExitRule$1.evaluate(SystemExitRule.java:39) app//org.junit.rules.RunRules.evaluate(RunRules.java:20) app//org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) app//org.junit.runners.ParentRunner.run(ParentRunner.java:413) app//org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:316) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:240) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:214) app//org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:155) app//org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:385) app//org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) app//org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:507) app//org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:495) Thread 2 (Reference Handler): State: RUNNABLE Blocked count: 3 Waited count: 0 Stack: java.base@17.0.11/java.lang.ref.Reference.waitForReferencePendingList(Native Method) java.base@17.0.11/java.lang.ref.Reference.processPendingReferences(Reference.java:253) java.base@17.0.11/java.lang.ref.Reference$ReferenceHandler.run(Reference.java:215) Thread 3 (Finalizer): State: WAITING Blocked count: 40 Waited count: 12 Waiting on java.lang.ref.ReferenceQueue$Lock@199c2830 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) java.base@17.0.11/java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:172) Thread 4 (Signal Dispatcher): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 12 (Common-Cleaner): State: TIMED_WAITING Blocked count: 18 Waited count: 20 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/jdk.internal.ref.CleanerImpl.run(CleanerImpl.java:140) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) java.base@17.0.11/jdk.internal.misc.InnocuousThread.run(InnocuousThread.java:162) Thread 13 (Notification Thread): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 14 (pool-1-thread-1): State: RUNNABLE Blocked count: 0 Waited count: 31 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.poll(EPollPort.java:200) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:281) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 15 (pool-1-thread-2): State: WAITING Blocked count: 0 Waited count: 29 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@31a5b7ac Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:275) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 16 (surefire-forkedjvm-stream-flusher): State: TIMED_WAITING Blocked count: 0 Waited count: 5776 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 18 (surefire-forkedjvm-command-thread): State: WAITING Blocked count: 0 Waited count: 59 Waiting on java.util.concurrent.CountDownLatch$Sync@60755c00 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.CountDownLatch.await(CountDownLatch.java:230) java.base@17.0.11/sun.nio.ch.PendingFuture.get(PendingFuture.java:178) app//org.apache.maven.surefire.api.util.internal.Channels$2.read(Channels.java:127) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:284) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) app//org.apache.maven.surefire.api.util.internal.Channels$3.readImpl(Channels.java:169) app//org.apache.maven.surefire.api.util.internal.AbstractNoninterruptibleReadableChannel.read(AbstractNoninterruptibleReadableChannel.java:50) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:430) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:419) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.readMessageType(AbstractStreamDecoder.java:116) app//org.apache.maven.surefire.booter.stream.CommandDecoder.decode(CommandDecoder.java:77) app//org.apache.maven.surefire.booter.spi.CommandChannelDecoder.decode(CommandChannelDecoder.java:60) app//org.apache.maven.surefire.booter.CommandReader$CommandRunnable.run(CommandReader.java:290) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 22 (Time-limited test): State: RUNNABLE Blocked count: 12870 Waited count: 13421 Stack: java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo1(Native Method) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:197) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:154) app//org.apache.hadoop.hbase.util.ReflectionUtils.printThreadInfo(ReflectionUtils.java:181) app//org.apache.hadoop.hbase.util.Threads.printThreadInfo(Threads.java:186) app//org.apache.hadoop.hbase.util.Threads.threadDumpingIsAlive(Threads.java:113) app//org.apache.hadoop.hbase.LocalHBaseCluster.join(LocalHBaseCluster.java:394) app//org.apache.hadoop.hbase.MiniHBaseCluster.waitUntilShutDown(MiniHBaseCluster.java:921) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniHBaseCluster(HBaseTestingUtility.java:1359) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniCluster(HBaseTestingUtility.java:1341) app//org.apache.hadoop.hbase.snapshot.TestExportSnapshot.tearDownAfterClass(TestExportSnapshot.java:121) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) java.base@17.0.11/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) java.base@17.0.11/java.lang.reflect.Method.invoke(Method.java:568) app//org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) app//org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) app//org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) app//org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) app//org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) Thread 23 (org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner): State: WAITING Blocked count: 12 Waited count: 13 Waiting on java.lang.ref.ReferenceQueue$Lock@2fdea105 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) app//org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner.run(FileSystem.java:4171) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 25 (SSL Certificates Store Monitor): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.TaskQueue@33da7b85 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 34 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@44d16179): State: TIMED_WAITING Blocked count: 0 Waited count: 1149 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 35 (GcTimeMonitor obsWindow = 60000, sleepInterval = 5000, maxGcTimePerc = 100): State: TIMED_WAITING Blocked count: 0 Waited count: 115 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.GcTimeMonitor.run(GcTimeMonitor.java:161) Thread 36 (pool-6-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 37 (qtp123845481-37): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 38 (qtp123845481-38): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 39 (qtp123845481-39): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 40 (qtp123845481-40): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 41 (qtp123845481-41-acceptor-0@781b1e42-ServerConnector@67ec1049{HTTP/1.1, (http/1.1)}{localhost:42647}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 42 (qtp123845481-42): State: TIMED_WAITING Blocked count: 0 Waited count: 10 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 43 (qtp123845481-43): State: TIMED_WAITING Blocked count: 0 Waited count: 10 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 44 (qtp123845481-44): State: TIMED_WAITING Blocked count: 0 Waited count: 10 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 45 (Session-HouseKeeper-604db391-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 46 (pool-7-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 53 (FSEditLogAsync): State: WAITING Blocked count: 28 Waited count: 3355 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@60a45dc1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.dequeueEdit(FSEditLogAsync.java:241) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.run(FSEditLogAsync.java:250) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 55 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 56 (IPC Server idle connection scanner for port 37495): State: TIMED_WAITING Blocked count: 1 Waited count: 59 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 58 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 115 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 61 (org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor@33e58e69): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor.run(PendingReconstructionBlocks.java:267) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 62 (DatanodeAdminMonitor-0): State: TIMED_WAITING Blocked count: 0 Waited count: 192 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 49 (org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor@989c5e0): State: TIMED_WAITING Blocked count: 0 Waited count: 115 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor.run(HeartbeatManager.java:563) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 47 (RedundancyMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 196 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) java.base@17.0.11/java.lang.Thread.sleep(Thread.java:344) java.base@17.0.11/java.util.concurrent.TimeUnit.sleep(TimeUnit.java:446) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$RedundancyMonitor.run(BlockManager.java:5352) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 48 (MarkedDeleteBlockScrubberThread): State: TIMED_WAITING Blocked count: 0 Waited count: 56703 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$MarkedDeleteBlockScrubber.run(BlockManager.java:5326) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 51 (Block report processor): State: WAITING Blocked count: 1 Waited count: 1625 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@468106b7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.processQueue(BlockManager.java:5627) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.run(BlockManager.java:5614) Thread 57 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 54 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 64 (IPC Server handler 0 on default port 37495): State: TIMED_WAITING Blocked count: 70 Waited count: 2562 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 65 (IPC Server handler 1 on default port 37495): State: TIMED_WAITING Blocked count: 70 Waited count: 2562 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 66 (IPC Server handler 2 on default port 37495): State: TIMED_WAITING Blocked count: 68 Waited count: 2567 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 67 (IPC Server handler 3 on default port 37495): State: TIMED_WAITING Blocked count: 66 Waited count: 2561 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 68 (IPC Server handler 4 on default port 37495): State: TIMED_WAITING Blocked count: 82 Waited count: 2550 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 69 (pool-12-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 71 (org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor@7d0a4f0): State: TIMED_WAITING Blocked count: 0 Waited count: 287 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor.run(LeaseManager.java:537) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 72 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor@108b955): State: TIMED_WAITING Blocked count: 0 Waited count: 115 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor.run(FSNamesystem.java:4550) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 73 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller@2ea8d10a): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller.run(FSNamesystem.java:4592) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 74 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber@5985a9b3): State: TIMED_WAITING Blocked count: 0 Waited count: 4 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber.run(FSNamesystem.java:4689) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 75 (CacheReplicationMonitor(822228812)): State: TIMED_WAITING Blocked count: 0 Waited count: 21 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1759) app//org.apache.hadoop.hdfs.server.blockmanagement.CacheReplicationMonitor.run(CacheReplicationMonitor.java:186) Thread 85 (pool-18-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 86 (qtp1135609267-86): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 87 (qtp1135609267-87-acceptor-0@473ea113-ServerConnector@3c5cb06a{HTTP/1.1, (http/1.1)}{localhost:34891}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 88 (qtp1135609267-88): State: TIMED_WAITING Blocked count: 0 Waited count: 10 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 89 (qtp1135609267-89): State: TIMED_WAITING Blocked count: 0 Waited count: 10 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 90 (Session-HouseKeeper-4ee4d1fd-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 91 (nioEventLoopGroup-2-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 92 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@3900f107): State: TIMED_WAITING Blocked count: 0 Waited count: 1145 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 94 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 95 (IPC Server idle connection scanner for port 45989): State: TIMED_WAITING Blocked count: 1 Waited count: 59 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 97 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 115 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 100 (Command processor): State: WAITING Blocked count: 3 Waited count: 336 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@18778abc Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 101 (BP-22138141-172.17.0.2-1736333101274 heartbeating to localhost/127.0.0.1:37495): State: TIMED_WAITING Blocked count: 1418 Waited count: 1616 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 102 (pool-20-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 84 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@8acd5ca): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 96 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 93 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 103 (IPC Server handler 0 on default port 45989): State: TIMED_WAITING Blocked count: 0 Waited count: 573 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 104 (IPC Server handler 1 on default port 45989): State: TIMED_WAITING Blocked count: 0 Waited count: 573 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 105 (IPC Server handler 2 on default port 45989): State: TIMED_WAITING Blocked count: 0 Waited count: 580 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 106 (IPC Server handler 3 on default port 45989): State: TIMED_WAITING Blocked count: 0 Waited count: 590 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 109 (IPC Server handler 4 on default port 45989): State: TIMED_WAITING Blocked count: 0 Waited count: 573 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 107 (IPC Client (175901922) connection to localhost/127.0.0.1:37495 from jenkins): State: TIMED_WAITING Blocked count: 1544 Waited count: 1545 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Thread 108 (IPC Parameter Sending Thread for localhost/127.0.0.1:37495): State: TIMED_WAITING Blocked count: 0 Waited count: 2299 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 119 (pool-26-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 120 (qtp40190603-120): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 121 (qtp40190603-121-acceptor-0@733970c9-ServerConnector@27b3056c{HTTP/1.1, (http/1.1)}{localhost:42931}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 122 (qtp40190603-122): State: TIMED_WAITING Blocked count: 0 Waited count: 10 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 123 (qtp40190603-123): State: TIMED_WAITING Blocked count: 0 Waited count: 10 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 124 (Session-HouseKeeper-66a6431-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 127 (nioEventLoopGroup-4-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 128 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@63d3adf1): State: TIMED_WAITING Blocked count: 0 Waited count: 1144 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 130 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 131 (IPC Server idle connection scanner for port 37919): State: TIMED_WAITING Blocked count: 1 Waited count: 59 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 133 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 115 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 136 (Command processor): State: WAITING Blocked count: 0 Waited count: 317 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@262fd62 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 137 (BP-22138141-172.17.0.2-1736333101274 heartbeating to localhost/127.0.0.1:37495): State: TIMED_WAITING Blocked count: 1440 Waited count: 1619 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 138 (pool-29-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 118 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@3e496c77): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 132 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 129 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 139 (IPC Server handler 0 on default port 37919): State: TIMED_WAITING Blocked count: 0 Waited count: 572 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 140 (IPC Server handler 1 on default port 37919): State: TIMED_WAITING Blocked count: 0 Waited count: 572 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 141 (IPC Server handler 2 on default port 37919): State: TIMED_WAITING Blocked count: 0 Waited count: 572 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 142 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data1)): State: TIMED_WAITING Blocked count: 7 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 144 (IPC Server handler 3 on default port 37919): State: TIMED_WAITING Blocked count: 0 Waited count: 572 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 146 (IPC Server handler 4 on default port 37919): State: TIMED_WAITING Blocked count: 0 Waited count: 572 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 145 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data2)): State: TIMED_WAITING Blocked count: 10 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 163 (pool-40-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 158 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data1/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 159 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data3)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 164 (qtp1264850999-164): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$254/0x00007f478c428988.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 165 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data4)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 166 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data2/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 167 (qtp1264850999-167-acceptor-0@758b1918-ServerConnector@3362f87c{HTTP/1.1, (http/1.1)}{localhost:40601}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 171 (qtp1264850999-171): State: TIMED_WAITING Blocked count: 0 Waited count: 10 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 172 (qtp1264850999-172): State: TIMED_WAITING Blocked count: 0 Waited count: 10 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 176 (Session-HouseKeeper-26f1d292-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 180 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data3/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 181 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data4/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 183 (pool-15-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 186 (java.util.concurrent.ThreadPoolExecutor$Worker@2c953ef0[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 190 (pool-23-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 193 (java.util.concurrent.ThreadPoolExecutor$Worker@1c56948b[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 194 (nioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 195 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@6eba8751): State: TIMED_WAITING Blocked count: 0 Waited count: 1143 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 197 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 198 (IPC Server idle connection scanner for port 45913): State: TIMED_WAITING Blocked count: 1 Waited count: 59 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 200 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 115 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 203 (Command processor): State: WAITING Blocked count: 1 Waited count: 355 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@336893f4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 204 (BP-22138141-172.17.0.2-1736333101274 heartbeating to localhost/127.0.0.1:37495): State: TIMED_WAITING Blocked count: 1405 Waited count: 1611 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 205 (pool-46-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 162 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@7b35f50f): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 199 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 196 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 206 (IPC Server handler 0 on default port 45913): State: TIMED_WAITING Blocked count: 0 Waited count: 573 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 207 (IPC Server handler 1 on default port 45913): State: TIMED_WAITING Blocked count: 0 Waited count: 572 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 208 (IPC Server handler 2 on default port 45913): State: TIMED_WAITING Blocked count: 0 Waited count: 572 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 209 (IPC Server handler 3 on default port 45913): State: TIMED_WAITING Blocked count: 0 Waited count: 573 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 210 (IPC Server handler 4 on default port 45913): State: TIMED_WAITING Blocked count: 0 Waited count: 576 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 213 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data5)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 214 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data6)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 220 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data5/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 2 Waited count: 3 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 221 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data6/current/BP-22138141-172.17.0.2-1736333101274): State: TIMED_WAITING Blocked count: 2 Waited count: 3 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 225 (pool-37-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 228 (java.util.concurrent.ThreadPoolExecutor$Worker@5fd6a7f8[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 233 (FsDatasetAsyncDiskServiceFixer): State: TIMED_WAITING Blocked count: 0 Waited count: 20 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer.run(HBaseTestingUtility.java:599) Thread 236 (NIOServerCxnFactory.SelectorThread-1): State: RUNNABLE Blocked count: 2 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 235 (NIOServerCxnFactory.SelectorThread-0): State: RUNNABLE Blocked count: 16 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 237 (NIOServerCxnFactory.AcceptThread:localhost/127.0.0.1:52367): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.select(NIOServerCnxnFactory.java:205) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.run(NIOServerCnxnFactory.java:181) Thread 234 (ConnnectionExpirer): State: TIMED_WAITING Blocked count: 0 Waited count: 58 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.NIOServerCnxnFactory$ConnectionExpirerThread.run(NIOServerCnxnFactory.java:554) Thread 238 (SessionTracker): State: TIMED_WAITING Blocked count: 0 Waited count: 286 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.SessionTrackerImpl.run(SessionTrackerImpl.java:163) Thread 239 (SyncThread:0): State: WAITING Blocked count: 33 Waited count: 728 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4dc1adaf Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.SyncRequestProcessor.run(SyncRequestProcessor.java:170) Thread 240 (ProcessThread(sid:0 cport:52367):): State: WAITING Blocked count: 0 Waited count: 818 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@d2bfd92 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.PrepRequestProcessor.run(PrepRequestProcessor.java:142) Thread 241 (RequestThrottler): State: WAITING Blocked count: 0 Waited count: 851 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@3c783c72 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.RequestThrottler.run(RequestThrottler.java:147) Thread 242 (NIOWorkerThread-1): State: WAITING Blocked count: 0 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 253 (Time-limited test.named-queue-events-pool-0): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@ff82e77 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) app//com.lmax.disruptor.BlockingWaitStrategy.waitFor(BlockingWaitStrategy.java:47) app//com.lmax.disruptor.ProcessingSequenceBarrier.waitFor(ProcessingSequenceBarrier.java:56) app//com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:159) app//com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 254 (HBase-Metrics2-1): State: TIMED_WAITING Blocked count: 0 Waited count: 376 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 255 (RS-EventLoopGroup-1-1): State: RUNNABLE Blocked count: 50 Waited count: 1 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 256 (Time-limited test-SendThread(127.0.0.1:52367)): State: RUNNABLE Blocked count: 20 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:332) app//org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1289) Thread 257 (Time-limited test-EventThread): State: WAITING Blocked count: 2 Waited count: 58 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@200ba914 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:550) Thread 258 (NIOWorkerThread-2): State: WAITING Blocked count: 3 Waited count: 138 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 259 (NIOWorkerThread-3): State: WAITING Blocked count: 3 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 260 (NIOWorkerThread-4): State: WAITING Blocked count: 1 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 261 (NIOWorkerThread-5): State: WAITING Blocked count: 3 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 262 (zk-event-processor-pool-0): State: WAITING Blocked count: 35 Waited count: 87 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4d05f8cc Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 263 (NIOWorkerThread-6): State: WAITING Blocked count: 4 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 264 (NIOWorkerThread-7): State: WAITING Blocked count: 3 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 265 (NIOWorkerThread-8): State: WAITING Blocked count: 0 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 266 (NIOWorkerThread-9): State: WAITING Blocked count: 2 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 267 (NIOWorkerThread-10): State: WAITING Blocked count: 3 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 268 (NIOWorkerThread-11): State: WAITING Blocked count: 6 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 269 (NIOWorkerThread-12): State: WAITING Blocked count: 1 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 270 (NIOWorkerThread-13): State: WAITING Blocked count: 2 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 271 (NIOWorkerThread-14): State: WAITING Blocked count: 3 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 272 (NIOWorkerThread-15): State: WAITING Blocked count: 2 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 273 (NIOWorkerThread-16): State: WAITING Blocked count: 2 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4f144eef Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 274 (RpcServer.default.FPBQ.Fifo.handler=0,queue=0,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@3c0ec8d Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 275 (RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=34551): State: WAITING Blocked count: 221 Waited count: 834 Waiting on java.util.concurrent.Semaphore$NonfairSync@32b787af Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 276 (RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34551): State: WAITING Blocked count: 93 Waited count: 356 Waiting on java.util.concurrent.Semaphore$NonfairSync@4f89cbe4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 277 (RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=34551): State: WAITING Blocked count: 80 Waited count: 5313 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@2d289941 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 278 (RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@2988bced Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 279 (RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@2988bced Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 280 (RpcServer.replication.FPBQ.Fifo.handler=0,queue=0,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@278ccc8c Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 281 (RpcServer.replication.FPBQ.Fifo.handler=1,queue=0,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@265be572 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 282 (RpcServer.replication.FPBQ.Fifo.handler=2,queue=0,port=34551): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@4fd08269 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 283 (RpcServer.metaPriority.FPBQ.Fifo.handler=0,queue=0,port=34551): State: WAITING Blocked count: 0 Waited count: 3 Waiting on java.util.concurrent.Semaphore$NonfairSync@23d320aa Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 287 (RS-EventLoopGroup-3-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 307 (RS-EventLoopGroup-4-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 327 (RS-EventLoopGroup-5-1): State: RUNNABLE Blocked count: 90 Waited count: 4 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 285 (M:0;67f42ee69b4e:34551): State: TIMED_WAITING Blocked count: 6 Waited count: 2683 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.FutureTask.awaitDone(FutureTask.java:444) java.base@17.0.11/java.util.concurrent.FutureTask.get(FutureTask.java:203) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.shutdown(AbstractFSWAL.java:1011) app//org.apache.hadoop.hbase.wal.AbstractFSWALProvider.shutdown(AbstractFSWALProvider.java:184) app//org.apache.hadoop.hbase.wal.WALFactory.shutdown(WALFactory.java:272) app//org.apache.hadoop.hbase.master.region.MasterRegion.shutdownWAL(MasterRegion.java:140) app//org.apache.hadoop.hbase.master.region.MasterRegion.close(MasterRegion.java:206) app//org.apache.hadoop.hbase.master.HMaster.stopServiceThreads(HMaster.java:1769) app//org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:1285) app//org.apache.hadoop.hbase.master.HMaster.run(HMaster.java:614) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 348 (Monitor thread for TaskMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 57 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.monitoring.TaskMonitor$MonitorRunnable.run(TaskMonitor.java:325) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 350 (master/67f42ee69b4e:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 352 (master/67f42ee69b4e:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 354 (org.apache.hadoop.hdfs.PeerCache@58490a7b): State: TIMED_WAITING Blocked count: 0 Waited count: 190 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.PeerCache.run(PeerCache.java:253) app//org.apache.hadoop.hdfs.PeerCache.access$000(PeerCache.java:46) app//org.apache.hadoop.hdfs.PeerCache$1.run(PeerCache.java:124) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 373 (master:store-WAL-Roller): State: TIMED_WAITING Blocked count: 0 Waited count: 5671 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hbase.wal.AbstractWALRoller.run(AbstractWALRoller.java:179) Thread 390 (RS-EventLoopGroup-5-2): State: RUNNABLE Blocked count: 108 Waited count: 3 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 391 (RS-EventLoopGroup-5-3): State: RUNNABLE Blocked count: 131 Waited count: 3 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 404 (Idle-Rpc-Conn-Sweeper-pool-0): State: WAITING Blocked count: 0 Waited count: 65 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6a97dfa0 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 415 (SnapshotHandlerChoreCleaner): State: TIMED_WAITING Blocked count: 0 Waited count: 57 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 403 (RpcClient-timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 56685 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 424 (RS-EventLoopGroup-1-2): State: RUNNABLE Blocked count: 27 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 425 (RS-EventLoopGroup-1-3): State: RUNNABLE Blocked count: 56 Waited count: 2 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 450 (RegionServerTracker-0): State: WAITING Blocked count: 8 Waited count: 10 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5be54308 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 472 (regionserver/67f42ee69b4e:0.procedureResultReporter): State: WAITING Blocked count: 14 Waited count: 29 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@508e73da Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 473 (regionserver/67f42ee69b4e:0.procedureResultReporter): State: WAITING Blocked count: 22 Waited count: 33 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@430dcb4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 474 (regionserver/67f42ee69b4e:0.procedureResultReporter): State: WAITING Blocked count: 14 Waited count: 23 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@595b70f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 512 (RS-EventLoopGroup-4-2): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 523 (RS-EventLoopGroup-4-3): State: RUNNABLE Blocked count: 2 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 525 (region-location-0): State: WAITING Blocked count: 8 Waited count: 12 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@741949b6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 551 (Async-Client-Retry-Timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 56459 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 566 (RPCClient-NioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 3 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 606 (region-location-1): State: WAITING Blocked count: 3 Waited count: 7 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@741949b6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 607 (region-location-2): State: WAITING Blocked count: 3 Waited count: 6 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@741949b6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 608 (region-location-3): State: WAITING Blocked count: 2 Waited count: 5 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@741949b6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1012 (MutableQuantiles-0): State: TIMED_WAITING Blocked count: 0 Waited count: 383 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1073 (RPCClient-NioEventLoopGroup-6-2): State: RUNNABLE Blocked count: 6 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1101 (RS-EventLoopGroup-3-2): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1112 (zk-permission-watcher-pool-0): State: WAITING Blocked count: 62 Waited count: 97 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7d78af0e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1120 (RPCClient-NioEventLoopGroup-6-3): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1121 (RS-EventLoopGroup-3-3): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1562 (Container metrics unregistration): State: WAITING Blocked count: 10 Waited count: 33 Waiting on java.util.TaskQueue@43921dee Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 2944 (RPCClient-NioEventLoopGroup-6-4): State: RUNNABLE Blocked count: 2 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 3542 (region-location-4): State: WAITING Blocked count: 3 Waited count: 6 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@741949b6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5203 (RPCClient-NioEventLoopGroup-6-5): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5204 (RPCClient-NioEventLoopGroup-6-6): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5205 (RPCClient-NioEventLoopGroup-6-7): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9395 (AsyncFSWAL-1-hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/MasterData-prefix:67f42ee69b4e,34551,1736333106957): State: WAITING Blocked count: 0 Waited count: 2 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@431a8944 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9399 (Timer for 'JobHistoryServer' metrics system): State: TIMED_WAITING Blocked count: 0 Waited count: 27 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 9400 (process reaper): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:401) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9403 (WAL-Shutdown-0): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.awaitTermination(ThreadPoolExecutor.java:1464) app//org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.doShutdown(AsyncFSWAL.java:793) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL$2.call(AbstractFSWAL.java:995) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL$2.call(AbstractFSWAL.java:990) java.base@17.0.11/java.util.concurrent.FutureTask.run(FutureTask.java:264) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9404 (Close-WAL-Writer-0): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:166) app//org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) app//org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.recoverLease(AsyncFSWAL.java:722) app//org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.lambda$closeWriter$5(AsyncFSWAL.java:735) app//org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL$$Lambda$1161/0x00007f478d15a120.run(Unknown Source) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) 2025-01-08T10:54:40,446 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(233): attempt=1 on file=hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/MasterData/WALs/67f42ee69b4e,34551,1736333106957/67f42ee69b4e%2C34551%2C1736333106957.1736333108999 after 4001ms java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.recoverLease(DFSClient.java:946) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:317) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:314) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.recoverLease(DistributedFileSystem.java:329) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverLease(RecoverLeaseFSUtils.java:222) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:155) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.recoverLease(AsyncFSWAL.java:722) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.lambda$closeWriter$5(AsyncFSWAL.java:735) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2025-01-08T10:54:41,440 ERROR [WAL-Shutdown-0 {}] wal.AsyncFSWAL(794): We have waited 5 seconds but the close of async writer doesn't complete.Please check the status of underlying filesystem or increase the wait time by the config "hbase.wal.async.wait.on.shutdown.seconds" 2025-01-08T10:54:41,440 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2025-01-08T10:54:41,440 INFO [M:0;67f42ee69b4e:34551 {}] flush.MasterFlushTableProcedureManager(91): stop: server shutting down. 2025-01-08T10:54:41,440 INFO [M:0;67f42ee69b4e:34551 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:34551 2025-01-08T10:54:41,443 DEBUG [M:0;67f42ee69b4e:34551 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/rs/67f42ee69b4e,34551,1736333106957 already deleted, retry=false 2025-01-08T10:54:41,448 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/MasterData/WALs/67f42ee69b4e,34551,1736333106957/67f42ee69b4e%2C34551%2C1736333106957.1736333108999 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.recoverLease(AsyncFSWAL.java:722) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.lambda$closeWriter$5(AsyncFSWAL.java:735) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 12 more 2025-01-08T10:54:41,546 INFO [M:0;67f42ee69b4e:34551 {}] regionserver.HRegionServer(1307): Exiting; stopping=67f42ee69b4e,34551,1736333106957; zookeeper connection closed. 2025-01-08T10:54:41,546 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2025-01-08T10:54:41,546 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34551-0x1000c1ee94b0000, quorum=127.0.0.1:52367, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2025-01-08T10:54:41,598 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@6cc6a12e{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2025-01-08T10:54:41,599 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@3362f87c{HTTP/1.1, (http/1.1)}{localhost:0} 2025-01-08T10:54:41,599 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2025-01-08T10:54:41,600 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@1e8b912e{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2025-01-08T10:54:41,601 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@29173d0{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop.log.dir/,STOPPED} 2025-01-08T10:54:41,605 WARN [BP-22138141-172.17.0.2-1736333101274 heartbeating to localhost/127.0.0.1:37495 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2025-01-08T10:54:41,605 WARN [BP-22138141-172.17.0.2-1736333101274 heartbeating to localhost/127.0.0.1:37495 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-22138141-172.17.0.2-1736333101274 (Datanode Uuid eceeeec2-0c7e-4597-a09d-1a300452c646) service to localhost/127.0.0.1:37495 2025-01-08T10:54:41,606 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2025-01-08T10:54:41,606 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2025-01-08T10:54:41,617 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data5/current/BP-22138141-172.17.0.2-1736333101274 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2025-01-08T10:54:41,623 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2025-01-08T10:54:41,626 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data6/current/BP-22138141-172.17.0.2-1736333101274 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2025-01-08T10:54:41,678 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@3abf71c7{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2025-01-08T10:54:41,679 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@27b3056c{HTTP/1.1, (http/1.1)}{localhost:0} 2025-01-08T10:54:41,679 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2025-01-08T10:54:41,679 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6c51149d{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2025-01-08T10:54:41,679 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@66733fff{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop.log.dir/,STOPPED} 2025-01-08T10:54:41,681 WARN [BP-22138141-172.17.0.2-1736333101274 heartbeating to localhost/127.0.0.1:37495 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2025-01-08T10:54:41,681 WARN [BP-22138141-172.17.0.2-1736333101274 heartbeating to localhost/127.0.0.1:37495 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-22138141-172.17.0.2-1736333101274 (Datanode Uuid b55dd1b9-f145-4b75-892c-153c95fea6e1) service to localhost/127.0.0.1:37495 2025-01-08T10:54:41,681 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2025-01-08T10:54:41,681 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2025-01-08T10:54:41,684 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data3/current/BP-22138141-172.17.0.2-1736333101274 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2025-01-08T10:54:41,684 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2025-01-08T10:54:41,686 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data4/current/BP-22138141-172.17.0.2-1736333101274 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2025-01-08T10:54:41,726 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@5a59e1d7{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2025-01-08T10:54:41,726 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@3c5cb06a{HTTP/1.1, (http/1.1)}{localhost:0} 2025-01-08T10:54:41,726 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2025-01-08T10:54:41,727 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@666781c4{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2025-01-08T10:54:41,727 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@617ad2bd{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop.log.dir/,STOPPED} 2025-01-08T10:54:41,729 WARN [BP-22138141-172.17.0.2-1736333101274 heartbeating to localhost/127.0.0.1:37495 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2025-01-08T10:54:41,729 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2025-01-08T10:54:41,729 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2025-01-08T10:54:41,729 WARN [BP-22138141-172.17.0.2-1736333101274 heartbeating to localhost/127.0.0.1:37495 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-22138141-172.17.0.2-1736333101274 (Datanode Uuid 2039b132-af90-4efa-b0f3-2607003b309d) service to localhost/127.0.0.1:37495 2025-01-08T10:54:41,732 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data1/current/BP-22138141-172.17.0.2-1736333101274 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2025-01-08T10:54:41,750 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/cluster_d665d950-4ae1-ecc4-96d4-410854bec204/dfs/data/data2/current/BP-22138141-172.17.0.2-1736333101274 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2025-01-08T10:54:41,750 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2025-01-08T10:54:41,778 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@146536f9{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2025-01-08T10:54:41,779 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@67ec1049{HTTP/1.1, (http/1.1)}{localhost:0} 2025-01-08T10:54:41,779 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2025-01-08T10:54:41,779 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@1ba6b2fb{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2025-01-08T10:54:41,779 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@d3238fd{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/e573e48e-fab4-da95-348c-080675052193/hadoop.log.dir/,STOPPED} 2025-01-08T10:54:41,998 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(345): Shutdown MiniZK cluster with all ZK servers 2025-01-08T10:54:42,450 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:37495/user/jenkins/test-data/52a3eaf0-1f9a-e272-1722-2e32972b316f/MasterData/WALs/67f42ee69b4e,34551,1736333106957/67f42ee69b4e%2C34551%2C1736333106957.1736333108999 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.recoverLease(AsyncFSWAL.java:722) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.lambda$closeWriter$5(AsyncFSWAL.java:735) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 12 more 2025-01-08T10:54:42,459 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1347): Minicluster is down