INFO [main] 2025-09-18 05:23:04,576 YamlConfigurationLoader.java:89 - Configuration location: file:/etc/cassandra/cassandra.yaml INFO [main] 2025-09-18 05:23:04,897 Config.java:495 - Node configuration:[allocate_tokens_for_keyspace=null; authenticator=AllowAllAuthenticator; authorizer=AllowAllAuthorizer; auto_bootstrap=true; auto_snapshot=true; back_pressure_enabled=false; back_pressure_strategy=org.apache.cassandra.net.RateBasedBackPressure{high_ratio=0.9, factor=5, flow=FAST}; batch_size_fail_threshold_in_kb=50; batch_size_warn_threshold_in_kb=5; batchlog_replay_throttle_in_kb=1024; broadcast_address=10.0.0.38; broadcast_rpc_address=10.0.0.38; buffer_pool_use_heap_if_exhausted=true; cas_contention_timeout_in_ms=1000; cdc_enabled=false; cdc_free_space_check_interval_ms=250; cdc_raw_directory=null; cdc_total_space_in_mb=0; client_encryption_options=; cluster_name=contrail_database; column_index_cache_size_in_kb=2; column_index_size_in_kb=64; commit_failure_policy=stop; commitlog_compression=null; commitlog_directory=/var/lib/cassandra/commitlog; commitlog_max_compression_buffers_in_pool=3; commitlog_periodic_queue_size=-1; commitlog_segment_size_in_mb=32; commitlog_sync=periodic; commitlog_sync_batch_window_in_ms=NaN; commitlog_sync_period_in_ms=10000; commitlog_total_space_in_mb=null; compaction_large_partition_warning_threshold_mb=100; compaction_throughput_mb_per_sec=256; concurrent_compactors=4; concurrent_counter_writes=32; concurrent_materialized_view_writes=32; concurrent_reads=64; concurrent_replicates=null; concurrent_writes=64; counter_cache_keys_to_save=2147483647; counter_cache_save_period=7200; counter_cache_size_in_mb=null; counter_write_request_timeout_in_ms=5000; credentials_cache_max_entries=1000; credentials_update_interval_in_ms=-1; credentials_validity_in_ms=2000; cross_node_timeout=false; data_file_directories=[Ljava.lang.String;@6b19b79; disk_access_mode=auto; disk_failure_policy=stop; disk_optimization_estimate_percentile=0.95; disk_optimization_page_cross_chance=0.1; disk_optimization_strategy=ssd; dynamic_snitch=true; dynamic_snitch_badness_threshold=0.1; dynamic_snitch_reset_interval_in_ms=600000; dynamic_snitch_update_interval_in_ms=100; enable_materialized_views=true; enable_scripted_user_defined_functions=false; enable_user_defined_functions=false; enable_user_defined_functions_threads=true; encryption_options=null; endpoint_snitch=SimpleSnitch; file_cache_round_up=null; file_cache_size_in_mb=null; gc_log_threshold_in_ms=200; gc_warn_threshold_in_ms=1000; hinted_handoff_disabled_datacenters=[]; hinted_handoff_enabled=true; hinted_handoff_throttle_in_kb=1024; hints_compression=null; hints_directory=null; hints_flush_period_in_ms=10000; incremental_backups=false; index_interval=null; index_summary_capacity_in_mb=null; index_summary_resize_interval_in_minutes=60; initial_token=null; inter_dc_stream_throughput_outbound_megabits_per_sec=200; inter_dc_tcp_nodelay=false; internode_authenticator=null; internode_compression=dc; internode_recv_buff_size_in_bytes=0; internode_send_buff_size_in_bytes=0; key_cache_keys_to_save=2147483647; key_cache_save_period=14400; key_cache_size_in_mb=null; listen_address=10.0.0.38; listen_interface=null; listen_interface_prefer_ipv6=false; listen_on_broadcast_address=false; max_hint_window_in_ms=10800000; max_hints_delivery_threads=2; max_hints_file_size_in_mb=128; max_mutation_size_in_kb=null; max_streaming_retries=3; max_value_size_in_mb=256; memtable_allocation_type=offheap_objects; memtable_cleanup_threshold=null; memtable_flush_writers=4; memtable_heap_space_in_mb=null; memtable_offheap_space_in_mb=null; min_free_space_per_drive_in_mb=50; native_transport_max_concurrent_connections=-1; native_transport_max_concurrent_connections_per_ip=-1; native_transport_max_frame_size_in_mb=256; native_transport_max_threads=128; native_transport_port=9042; native_transport_port_ssl=null; num_tokens=256; otc_backlog_expiration_interval_ms=200; otc_coalescing_enough_coalesced_messages=8; otc_coalescing_strategy=DISABLED; otc_coalescing_window_us=200; partitioner=org.apache.cassandra.dht.Murmur3Partitioner; permissions_cache_max_entries=1000; permissions_update_interval_in_ms=-1; permissions_validity_in_ms=2000; phi_convict_threshold=8.0; prepared_statements_cache_size_mb=null; range_request_timeout_in_ms=10000; read_request_timeout_in_ms=5000; request_scheduler=org.apache.cassandra.scheduler.NoScheduler; request_scheduler_id=null; request_scheduler_options=null; request_timeout_in_ms=10000; role_manager=CassandraRoleManager; roles_cache_max_entries=1000; roles_update_interval_in_ms=-1; roles_validity_in_ms=2000; row_cache_class_name=org.apache.cassandra.cache.OHCProvider; row_cache_keys_to_save=2147483647; row_cache_save_period=0; row_cache_size_in_mb=0; rpc_address=10.0.0.38; rpc_interface=null; rpc_interface_prefer_ipv6=false; rpc_keepalive=true; rpc_listen_backlog=50; rpc_max_threads=2147483647; rpc_min_threads=16; rpc_port=9160; rpc_recv_buff_size_in_bytes=null; rpc_send_buff_size_in_bytes=null; rpc_server_type=sync; saved_caches_directory=/var/lib/cassandra/saved_caches; seed_provider=org.apache.cassandra.locator.SimpleSeedProvider{seeds=10.0.0.22,10.0.0.38}; server_encryption_options=; slow_query_log_timeout_in_ms=500; snapshot_before_compaction=false; ssl_storage_port=7001; sstable_preemptive_open_interval_in_mb=50; start_native_transport=true; start_rpc=true; storage_port=7000; stream_throughput_outbound_megabits_per_sec=200; streaming_keep_alive_period_in_secs=300; streaming_socket_timeout_in_ms=86400000; thrift_framed_transport_size_in_mb=15; thrift_max_message_length_in_mb=16; thrift_prepared_statements_cache_size_mb=null; tombstone_failure_threshold=100000; tombstone_warn_threshold=1000; tracetype_query_ttl=86400; tracetype_repair_ttl=604800; transparent_data_encryption_options=org.apache.cassandra.config.TransparentDataEncryptionOptions@2a32de6c; trickle_fsync=false; trickle_fsync_interval_in_kb=10240; truncate_request_timeout_in_ms=60000; unlogged_batch_across_partitions_warn_threshold=10; user_defined_function_fail_timeout=1500; user_defined_function_warn_timeout=500; user_function_timeout_policy=die; windows_timer_interval=1; write_request_timeout_in_ms=2000] INFO [main] 2025-09-18 05:23:04,898 DatabaseDescriptor.java:367 - DiskAccessMode 'auto' determined to be mmap, indexAccessMode is mmap INFO [main] 2025-09-18 05:23:04,899 DatabaseDescriptor.java:425 - Global memtable on-heap threshold is enabled at 502MB INFO [main] 2025-09-18 05:23:04,899 DatabaseDescriptor.java:429 - Global memtable off-heap threshold is enabled at 502MB INFO [main] 2025-09-18 05:23:04,927 RateBasedBackPressure.java:123 - Initialized back-pressure with high ratio: 0.9, factor: 5, flow: FAST, window size: 2000. INFO [main] 2025-09-18 05:23:04,928 DatabaseDescriptor.java:729 - Back-pressure is disabled with strategy org.apache.cassandra.net.RateBasedBackPressure{high_ratio=0.9, factor=5, flow=FAST}. INFO [main] 2025-09-18 05:23:05,147 JMXServerUtils.java:246 - Configured JMX server at: service:jmx:rmi://0.0.0.0/jndi/rmi://0.0.0.0:7201/jmxrmi INFO [main] 2025-09-18 05:23:05,153 CassandraDaemon.java:473 - Hostname: cn-jenkins-deploy-platform-ansible-os-4134-2. INFO [main] 2025-09-18 05:23:05,153 CassandraDaemon.java:480 - JVM vendor/version: OpenJDK 64-Bit Server VM/1.8.0_322 INFO [main] 2025-09-18 05:23:05,161 CassandraDaemon.java:481 - Heap size: 984.000MiB/1.961GiB INFO [main] 2025-09-18 05:23:05,164 CassandraDaemon.java:486 - Code Cache Non-heap memory: init = 2555904(2496K) used = 4350272(4248K) committed = 4390912(4288K) max = 251658240(245760K) INFO [main] 2025-09-18 05:23:05,165 CassandraDaemon.java:486 - Metaspace Non-heap memory: init = 0(0K) used = 19217032(18766K) committed = 19660800(19200K) max = -1(-1K) INFO [main] 2025-09-18 05:23:05,165 CassandraDaemon.java:486 - Compressed Class Space Non-heap memory: init = 0(0K) used = 2236160(2183K) committed = 2359296(2304K) max = 1073741824(1048576K) INFO [main] 2025-09-18 05:23:05,165 CassandraDaemon.java:486 - Par Eden Space Heap memory: init = 335544320(327680K) used = 93992528(91789K) committed = 335544320(327680K) max = 335544320(327680K) INFO [main] 2025-09-18 05:23:05,165 CassandraDaemon.java:486 - Par Survivor Space Heap memory: init = 41943040(40960K) used = 0(0K) committed = 41943040(40960K) max = 41943040(40960K) INFO [main] 2025-09-18 05:23:05,176 CassandraDaemon.java:486 - CMS Old Gen Heap memory: init = 654311424(638976K) used = 0(0K) committed = 654311424(638976K) max = 1728053248(1687552K) INFO [main] 2025-09-18 05:23:05,176 CassandraDaemon.java:488 - Classpath: /opt/cassandra/conf:/opt/cassandra/build/classes/main:/opt/cassandra/build/classes/thrift:/opt/cassandra/lib/airline-0.6.jar:/opt/cassandra/lib/antlr-runtime-3.5.2.jar:/opt/cassandra/lib/apache-cassandra-3.11.3.jar:/opt/cassandra/lib/apache-cassandra-thrift-3.11.3.jar:/opt/cassandra/lib/asm-5.0.4.jar:/opt/cassandra/lib/caffeine-2.2.6.jar:/opt/cassandra/lib/cassandra-driver-core-3.0.1-shaded.jar:/opt/cassandra/lib/commons-cli-1.1.jar:/opt/cassandra/lib/commons-codec-1.9.jar:/opt/cassandra/lib/commons-lang3-3.1.jar:/opt/cassandra/lib/commons-math3-3.2.jar:/opt/cassandra/lib/compress-lzf-0.8.4.jar:/opt/cassandra/lib/concurrentlinkedhashmap-lru-1.4.jar:/opt/cassandra/lib/concurrent-trees-2.4.0.jar:/opt/cassandra/lib/disruptor-3.0.1.jar:/opt/cassandra/lib/ecj-4.4.2.jar:/opt/cassandra/lib/guava-18.0.jar:/opt/cassandra/lib/HdrHistogram-2.1.9.jar:/opt/cassandra/lib/high-scale-lib-1.0.6.jar:/opt/cassandra/lib/hppc-0.5.4.jar:/opt/cassandra/lib/jackson-core-asl-1.9.13.jar:/opt/cassandra/lib/jackson-mapper-asl-1.9.13.jar:/opt/cassandra/lib/jamm-0.3.0.jar:/opt/cassandra/lib/javax.inject.jar:/opt/cassandra/lib/jbcrypt-0.3m.jar:/opt/cassandra/lib/jcl-over-slf4j-1.7.7.jar:/opt/cassandra/lib/jctools-core-1.2.1.jar:/opt/cassandra/lib/jflex-1.6.0.jar:/opt/cassandra/lib/jna-4.2.2.jar:/opt/cassandra/lib/joda-time-2.4.jar:/opt/cassandra/lib/json-simple-1.1.jar:/opt/cassandra/lib/jstackjunit-0.0.1.jar:/opt/cassandra/lib/libthrift-0.13.0.jar:/opt/cassandra/lib/log4j-over-slf4j-1.7.7.jar:/opt/cassandra/lib/logback-classic-1.2.9.jar:/opt/cassandra/lib/logback-core-1.2.9.jar:/opt/cassandra/lib/lz4-1.3.0.jar:/opt/cassandra/lib/metrics-core-3.1.5.jar:/opt/cassandra/lib/metrics-jvm-3.1.5.jar:/opt/cassandra/lib/metrics-logback-3.1.5.jar:/opt/cassandra/lib/netty-all-4.1.39.Final.jar:/opt/cassandra/lib/ohc-core-0.4.4.jar:/opt/cassandra/lib/ohc-core-j8-0.4.4.jar:/opt/cassandra/lib/reporter-config3-3.0.3.jar:/opt/cassandra/lib/reporter-config-base-3.0.3.jar:/opt/cassandra/lib/sigar-1.6.4.jar:/opt/cassandra/lib/slf4j-api-1.7.7.jar:/opt/cassandra/lib/snakeyaml-1.11.jar:/opt/cassandra/lib/snappy-java-1.1.1.7.jar:/opt/cassandra/lib/snowball-stemmer-1.3.0.581.1.jar:/opt/cassandra/lib/ST4-4.0.8.jar:/opt/cassandra/lib/stream-2.5.2.jar:/opt/cassandra/lib/thrift-server-0.3.7.jar:/opt/cassandra/lib/jsr223/*/*.jar:/opt/cassandra/lib/jamm-0.3.0.jar INFO [main] 2025-09-18 05:23:05,177 CassandraDaemon.java:490 - JVM Arguments: [-Xloggc:/opt/cassandra/logs/gc.log, -ea, -XX:+UseThreadPriorities, -XX:ThreadPriorityPolicy=42, -XX:+HeapDumpOnOutOfMemoryError, -Xss256k, -XX:StringTableSize=1000003, -XX:+AlwaysPreTouch, -XX:-UseBiasedLocking, -XX:+UseTLAB, -XX:+ResizeTLAB, -XX:+UseNUMA, -XX:+PerfDisableSharedMem, -Djava.net.preferIPv4Stack=true, -Xms1g, -Xmx2g, -XX:+UseParNewGC, -XX:+UseConcMarkSweepGC, -XX:+CMSParallelRemarkEnabled, -XX:SurvivorRatio=8, -XX:MaxTenuringThreshold=1, -XX:CMSInitiatingOccupancyFraction=75, -XX:+UseCMSInitiatingOccupancyOnly, -XX:CMSWaitDuration=10000, -XX:+CMSParallelInitialMarkEnabled, -XX:+CMSEdenChunksRecordAlways, -XX:+CMSClassUnloadingEnabled, -XX:+PrintGCDetails, -XX:+PrintGCDateStamps, -XX:+PrintHeapAtGC, -XX:+PrintTenuringDistribution, -XX:+PrintGCApplicationStoppedTime, -XX:+PrintPromotionFailure, -XX:+UseGCLogFileRotation, -XX:NumberOfGCLogFiles=10, -XX:GCLogFileSize=10M, -Xmn400M, -XX:+UseCondCardMark, -XX:CompileCommandFile=/opt/cassandra/conf/hotspot_compiler, -javaagent:/opt/cassandra/lib/jamm-0.3.0.jar, -Dcassandra.jmx.remote.port=7199, -Dcom.sun.management.jmxremote.rmi.port=7199, -Dcom.sun.management.jmxremote.authenticate=true, -Dcom.sun.management.jmxremote.password.file=/etc/cassandra/jmxremote.password, -Djava.library.path=/opt/cassandra/lib/sigar-bin, -Dcassandra.rpc_port=9161, -Dcassandra.native_transport_port=9041, -Dcassandra.ssl_storage_port=7013, -Dcassandra.storage_port=7012, -Dcassandra.jmx.local.port=7201, -Dcom.sun.management.jmxremote.access.file=/etc/cassandra/jmxremote.access, -Dcassandra.jmx.remote.port=7201, -Dcom.sun.management.jmxremote.rmi.port=7201, -Dcassandra.libjemalloc=/usr/lib64/libjemalloc.so.1, -XX:OnOutOfMemoryError=kill -9 %p, -Dlogback.configurationFile=logback.xml, -Dcassandra.logdir=/opt/cassandra/logs, -Dcassandra.storagedir=/opt/cassandra/data, -Dcassandra-foreground=yes] WARN [main] 2025-09-18 05:23:05,284 NativeLibrary.java:187 - Unable to lock JVM memory (ENOMEM). This can result in part of the JVM being swapped out, especially with mmapped I/O enabled. Increase RLIMIT_MEMLOCK or run Cassandra as root. INFO [main] 2025-09-18 05:23:05,284 StartupChecks.java:140 - jemalloc seems to be preloaded from /usr/lib64/libjemalloc.so.1 INFO [main] 2025-09-18 05:23:05,284 StartupChecks.java:176 - JMX is enabled to receive remote connections on port: 7201 INFO [main] 2025-09-18 05:23:05,286 SigarLibrary.java:44 - Initializing SIGAR library INFO [main] 2025-09-18 05:23:05,338 SigarLibrary.java:180 - Checked OS settings and found them configured for optimal performance. WARN [main] 2025-09-18 05:23:05,339 StartupChecks.java:311 - Maximum number of memory map areas per process (vm.max_map_count) 128960 is too low, recommended value: 1048575, you can change it with sysctl. WARN [main] 2025-09-18 05:23:05,364 StartupChecks.java:332 - Directory /var/lib/cassandra/commitlog doesn't exist WARN [main] 2025-09-18 05:23:05,366 StartupChecks.java:332 - Directory /var/lib/cassandra/saved_caches doesn't exist WARN [main] 2025-09-18 05:23:05,370 StartupChecks.java:332 - Directory /opt/cassandra/data/hints doesn't exist INFO [main] 2025-09-18 05:23:05,423 QueryProcessor.java:116 - Initialized prepared statement caches with 10 MB (native) and 10 MB (Thrift) INFO [main] 2025-09-18 05:23:06,009 ColumnFamilyStore.java:411 - Initializing system.IndexInfo INFO [main] 2025-09-18 05:23:07,214 ColumnFamilyStore.java:411 - Initializing system.batches INFO [main] 2025-09-18 05:23:07,240 ColumnFamilyStore.java:411 - Initializing system.paxos INFO [main] 2025-09-18 05:23:07,271 ColumnFamilyStore.java:411 - Initializing system.local INFO [main] 2025-09-18 05:23:07,307 ColumnFamilyStore.java:411 - Initializing system.peers INFO [main] 2025-09-18 05:23:07,319 ColumnFamilyStore.java:411 - Initializing system.peer_events INFO [main] 2025-09-18 05:23:07,335 ColumnFamilyStore.java:411 - Initializing system.range_xfers INFO [main] 2025-09-18 05:23:07,355 ColumnFamilyStore.java:411 - Initializing system.compaction_history INFO [main] 2025-09-18 05:23:07,361 ColumnFamilyStore.java:411 - Initializing system.sstable_activity INFO [main] 2025-09-18 05:23:07,384 ColumnFamilyStore.java:411 - Initializing system.size_estimates INFO [main] 2025-09-18 05:23:07,397 ColumnFamilyStore.java:411 - Initializing system.available_ranges INFO [main] 2025-09-18 05:23:07,418 ColumnFamilyStore.java:411 - Initializing system.transferred_ranges INFO [main] 2025-09-18 05:23:07,432 ColumnFamilyStore.java:411 - Initializing system.views_builds_in_progress INFO [main] 2025-09-18 05:23:07,449 ColumnFamilyStore.java:411 - Initializing system.built_views INFO [main] 2025-09-18 05:23:07,472 ColumnFamilyStore.java:411 - Initializing system.hints INFO [main] 2025-09-18 05:23:07,514 ColumnFamilyStore.java:411 - Initializing system.batchlog INFO [main] 2025-09-18 05:23:07,534 ColumnFamilyStore.java:411 - Initializing system.prepared_statements INFO [main] 2025-09-18 05:23:07,539 ColumnFamilyStore.java:411 - Initializing system.schema_keyspaces INFO [main] 2025-09-18 05:23:07,564 ColumnFamilyStore.java:411 - Initializing system.schema_columnfamilies INFO [main] 2025-09-18 05:23:07,582 ColumnFamilyStore.java:411 - Initializing system.schema_columns INFO [main] 2025-09-18 05:23:07,595 ColumnFamilyStore.java:411 - Initializing system.schema_triggers INFO [main] 2025-09-18 05:23:07,620 ColumnFamilyStore.java:411 - Initializing system.schema_usertypes INFO [main] 2025-09-18 05:23:07,634 ColumnFamilyStore.java:411 - Initializing system.schema_functions INFO [main] 2025-09-18 05:23:07,648 ColumnFamilyStore.java:411 - Initializing system.schema_aggregates INFO [main] 2025-09-18 05:23:07,653 ViewManager.java:137 - Not submitting build tasks for views in keyspace system as storage service is not initialized INFO [main] 2025-09-18 05:23:07,994 ApproximateTime.java:44 - Scheduling approximate time-check task with a precision of 10 milliseconds INFO [main] 2025-09-18 05:23:08,074 ColumnFamilyStore.java:411 - Initializing system_schema.keyspaces INFO [main] 2025-09-18 05:23:08,087 ColumnFamilyStore.java:411 - Initializing system_schema.tables INFO [main] 2025-09-18 05:23:08,160 ColumnFamilyStore.java:411 - Initializing system_schema.columns INFO [main] 2025-09-18 05:23:08,176 ColumnFamilyStore.java:411 - Initializing system_schema.triggers INFO [main] 2025-09-18 05:23:08,194 ColumnFamilyStore.java:411 - Initializing system_schema.dropped_columns INFO [main] 2025-09-18 05:23:08,211 ColumnFamilyStore.java:411 - Initializing system_schema.views INFO [main] 2025-09-18 05:23:08,235 ColumnFamilyStore.java:411 - Initializing system_schema.types INFO [main] 2025-09-18 05:23:08,244 ColumnFamilyStore.java:411 - Initializing system_schema.functions INFO [main] 2025-09-18 05:23:08,279 ColumnFamilyStore.java:411 - Initializing system_schema.aggregates INFO [main] 2025-09-18 05:23:08,287 ColumnFamilyStore.java:411 - Initializing system_schema.indexes INFO [main] 2025-09-18 05:23:08,288 ViewManager.java:137 - Not submitting build tasks for views in keyspace system_schema as storage service is not initialized INFO [MemtableFlushWriter:1] 2025-09-18 05:23:09,187 CacheService.java:112 - Initializing key cache with capacity of 49 MBs. INFO [MemtableFlushWriter:1] 2025-09-18 05:23:09,197 CacheService.java:134 - Initializing row cache with capacity of 0 MBs INFO [MemtableFlushWriter:1] 2025-09-18 05:23:09,200 CacheService.java:163 - Initializing counter cache with capacity of 24 MBs INFO [MemtableFlushWriter:1] 2025-09-18 05:23:09,200 CacheService.java:174 - Scheduling counter cache save to every 7200 seconds (going to save all keys). INFO [CompactionExecutor:4] 2025-09-18 05:23:09,582 BufferPool.java:230 - Global buffer pool is enabled, when pool is exhausted (max is 502.000MiB) it will allocate on heap INFO [main] 2025-09-18 05:23:09,744 StorageService.java:600 - Populating token metadata from system tables INFO [main] 2025-09-18 05:23:09,935 StorageService.java:607 - Token metadata: INFO [pool-4-thread-1] 2025-09-18 05:23:10,030 AutoSavingCache.java:174 - Completed loading (2 ms; 1 keys) KeyCache cache INFO [main] 2025-09-18 05:23:10,067 CommitLog.java:152 - No commitlog files found; skipping replay INFO [main] 2025-09-18 05:23:10,067 StorageService.java:600 - Populating token metadata from system tables INFO [main] 2025-09-18 05:23:10,143 StorageService.java:607 - Token metadata: INFO [main] 2025-09-18 05:23:10,295 QueryProcessor.java:163 - Preloaded 0 prepared statements INFO [main] 2025-09-18 05:23:10,297 StorageService.java:618 - Cassandra version: 3.11.3 INFO [main] 2025-09-18 05:23:10,299 StorageService.java:619 - Thrift API version: 20.1.0 INFO [main] 2025-09-18 05:23:10,303 StorageService.java:620 - CQL supported versions: 3.4.4 (default: 3.4.4) INFO [main] 2025-09-18 05:23:10,303 StorageService.java:622 - Native protocol supported versions: 3/v3, 4/v4, 5/v5-beta (default: 4/v4) INFO [main] 2025-09-18 05:23:10,360 IndexSummaryManager.java:85 - Initializing index summary manager with a memory pool size of 49 MB and a resize interval of 60 minutes INFO [main] 2025-09-18 05:23:10,382 MessagingService.java:761 - Starting Messaging Service on /10.0.0.38:7012 (ens3) WARN [main] 2025-09-18 05:23:10,399 SystemKeyspace.java:1087 - No host ID found, created 198ff7cf-da95-458a-aeb6-e3637ec669aa (Note: This should happen exactly once per node). INFO [main] 2025-09-18 05:23:10,462 OutboundTcpConnection.java:108 - OutboundTcpConnection using coalescing strategy DISABLED INFO [HANDSHAKE-/10.0.0.33] 2025-09-18 05:23:10,579 OutboundTcpConnection.java:561 - Handshaking version with /10.0.0.33 INFO [HANDSHAKE-/10.0.0.22] 2025-09-18 05:23:11,630 OutboundTcpConnection.java:561 - Handshaking version with /10.0.0.22 INFO [main] 2025-09-18 05:23:12,520 StorageService.java:550 - Unable to gossip with any peers but continuing anyway since node is in its own seed list INFO [main] 2025-09-18 05:23:12,544 StorageService.java:704 - Loading persisted ring state INFO [main] 2025-09-18 05:23:12,545 StorageService.java:822 - Starting up server gossip INFO [main] 2025-09-18 05:23:12,657 StorageService.java:883 - This node will not auto bootstrap because it is configured to be a seed node. INFO [main] 2025-09-18 05:23:12,674 BootStrapper.java:228 - Generated random tokens. tokens are [964485437255120294, -5932377793872121683, -5410308143724540828, -3047781528472203520, -3205217788789494885, 6210684016986443425, -7586511266047119285, -757408861089503534, -351705973192821726, -1300582166746074256, -4970139577644743915, 3245648289248929919, -3881458773210580086, -2898467036911510537, 1127384304222581043, -3836110817800358752, -2914482981178394045, 2117328061304225502, -7270576858263334702, 6696478556808590255, 4440569933823317634, 8454928253213658420, 4667909758121318836, -2716750045221122751, 7256511931855943447, 3267030986733485142, 8127616069451618638, -1729368167216901685, 5461182478799599891, 3340576315269385898, -8235735209558572574, 738637311497449832, -7384943639680539176, -2560509990857338135, -7574499572300845130, 3149809850129178704, -4779683873682396863, -7008578145743271460, -4846631810263962399, -7906812338217623354, -3706560632678465858, 1271601900459964512, -6286920548717944552, -69708076395865602, 881626031675052269, -3533030446991337648, -8752694935280654464, 4818584835403758988, -2882419553896237716, -1621403448561396264, 7038392457546354357, 3501469055805294138, 1790044107659659854, -3183211775600080646, -6924089855226659547, 3266263177841904641, -658819807063036083, 7941367473979278978, 6114418130054945384, 6928028511761453624, 7740021582969549633, 75695366645868515, 8365445482787888463, 6224211044271524403, -693552898453123192, -3767975118296647475, -7574071686978321448, 4457865785905849666, 3830744707988958024, -8391614153872508884, 8799753400187703472, 2125637406720438977, -7259066200253313333, -3293498149522017429, -2620927030340547666, -5011890912026858867, 4143655380372720455, -2795998571896326124, -5601671449566501444, 8481024201892278464, 7427326461656260165, 8389064290871951379, 36476571826789707, 518814943348996243, 5774383900335976794, 6508985318858402719, 1114242788254216026, -7612481605053330119, 7280403020131444416, 1596801994008150803, 1433204480358917885, -5955355394073103570, 7872425199870947806, 853661768332462199, -3013039769601996700, -1807259092605306096, -3108043403685452234, 5659993337038223490, 3924546623166558375, 4562215864280406444, -6216834833033093858, -3374948263785175475, -5517623686363677731, 9063466815532405597, -3750463226273909643, 3166095761252567076, -7175717594401518258, 3336083577724914561, 7983196646232853470, 8271182323744512982, 884832867448362892, -4135559629646352048, -7333941025793667458, -52736900454802333, 4885110719842166238, -1830959951466861773, -8975666107877808190, -1423407625409017229, 1036931925837975885, 1102292207352698556, -363605146416477115, -6586184957951168106, 359897270587778465, 154514073754459010, 2968650428472413055, 7005144239581530864, 7393983484019645629, -6601336222208555154, 1671018064282221850, 7902604814337951884, 5582944990751521062, 3446815279945612437, -7719571434659436972, -7486331036231874788, -8481477160987180806, -4275856648375796813, 975019040492789234, 1545109410920203684, -4183756986871169754, -1427865403170138070, -78319228150336010, -9134516030907875651, -7546015284750382192, 6207841748676171179, -3846769464494905880, -8603850841133973132, -2821514627815557923, 8465734271561408348, -2108583828829684993, 7494415186461120433, 4491427750528076413, -1479209432064143407, -5457160231678556519, 7177802933260429699, -4056815216463396881, -1423560939911900784, -1972507862853621505, -4072002462313952177, -8173988711775076942, -5861834118414866177, -6283904446036533588, 1922951116883820553, 2256011305137464276, -1260541438162455519, -1472053536352248605, 7929101885032422949, 1390503002539491433, 4976129964292817989, -7051510370892492492, 4281345940153206842, 6272997560524457393, 1276720718995698523, 799352137055018827, 1287154753277825775, 1056034994022223556, -4253692787788461709, -6362734777048780952, -7922339663332442024, 828295324755690294, -4406767420158546877, -5519258933237726726, 7945637447647517994, 2811525628858781426, 6018591951904543460, -8020449262599139914, -7690941618674222624, -6225026018770312521, -8070058689753124981, 6053395530787893189, 5973353265293454622, -137313085624354979, -5282825498978313988, -5250275963286563819, -7055033307230369263, -4249546593904607668, 3292453629067344866, 6999569276460355925, 9008053231639129969, -4739787876725351854, 7271133474648053784, -4849834365226946121, -1287938692963076422, -8454148968884885764, 6123439218397768887, 6070571161624821650, 3794781568720267103, 7724030247323064382, -748028049694524023, -6576523930440582051, -7831689022682190535, -863507992324021424, -9083771317662873172, -4546457910950102428, 1963667213076878527, -2405810815761716365, -1909327303462251554, -5590341154564265133, 3505324523750115768, 3052604082748987625, 404480391789260751, -204440653719437918, -9064397874770026951, 543853034383062838, -4191904565443883243, -4001846181061434915, -6262210575197844300, -1307004358803252316, -6659379084563805744, 9114596560563111895, -1845804463445041976, 5228052482064703906, 6521650935229633740, -9209246771099019622, -622987463617473748, -1795788871333808166, -4151132058289080004, 3700756026632897724, -2873360351589131805, 1084461684206485337, 6254665573363841308, 603007711604466385, 6370013610491411357, -3788247282668580207, -4223731180425976402, -4228214596932855234, -8011510321982256855, 4040204749760298086, 5746988048506612898, 7794479627789419843, 678101253642175213, 7399330528829345851, -882371362824440576, 5482710683617896693, -5083698599490360897, -4760446837840344923, 1169695687312442961] INFO [main] 2025-09-18 05:23:12,682 MigrationManager.java:331 - Create new Keyspace: KeyspaceMetadata{name=system_traces, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=2}}, tables=[org.apache.cassandra.config.CFMetaData@29deed7b[cfId=c5e99f16-8677-3914-b17e-960613512345,ksName=system_traces,cfName=sessions,flags=[COMPOUND],params=TableParams{comment=tracing sessions, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=0, default_time_to_live=0, memtable_flush_period_in_ms=3600000, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [client command coordinator duration request started_at parameters]],partitionKeyColumns=[session_id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UUIDType,columnMetadata=[client, command, session_id, coordinator, request, started_at, duration, parameters],droppedColumns={},triggers=[],indexes=[]], org.apache.cassandra.config.CFMetaData@310e6c61[cfId=8826e8e9-e16a-3728-8753-3bc1fc713c25,ksName=system_traces,cfName=events,flags=[COMPOUND],params=TableParams{comment=tracing events, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=0, default_time_to_live=0, memtable_flush_period_in_ms=3600000, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[] | [activity source source_elapsed thread]],partitionKeyColumns=[session_id],clusteringColumns=[event_id],keyValidator=org.apache.cassandra.db.marshal.UUIDType,columnMetadata=[activity, event_id, session_id, source, thread, source_elapsed],droppedColumns={},triggers=[],indexes=[]]], views=[], functions=[], types=[]} INFO [MigrationStage:1] 2025-09-18 05:23:13,279 ViewManager.java:137 - Not submitting build tasks for views in keyspace system_traces as storage service is not initialized INFO [MigrationStage:1] 2025-09-18 05:23:13,326 ColumnFamilyStore.java:411 - Initializing system_traces.events INFO [MigrationStage:1] 2025-09-18 05:23:13,361 ColumnFamilyStore.java:411 - Initializing system_traces.sessions INFO [main] 2025-09-18 05:23:13,394 MigrationManager.java:331 - Create new Keyspace: KeyspaceMetadata{name=system_distributed, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=3}}, tables=[org.apache.cassandra.config.CFMetaData@7e6a5b7e[cfId=759fffad-624b-3181-80ee-fa9a52d1f627,ksName=system_distributed,cfName=repair_history,flags=[COMPOUND],params=TableParams{comment=Repair history, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=3600000, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[] | [coordinator exception_message exception_stacktrace finished_at parent_id range_begin range_end started_at status participants]],partitionKeyColumns=[keyspace_name, columnfamily_name],clusteringColumns=[id],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[status, id, coordinator, finished_at, participants, exception_stacktrace, parent_id, range_end, range_begin, exception_message, keyspace_name, started_at, columnfamily_name],droppedColumns={},triggers=[],indexes=[]], org.apache.cassandra.config.CFMetaData@4f08a11[cfId=deabd734-b99d-3b9c-92e5-fd92eb5abf14,ksName=system_distributed,cfName=parent_repair_history,flags=[COMPOUND],params=TableParams{comment=Repair history, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=3600000, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [exception_message exception_stacktrace finished_at keyspace_name started_at columnfamily_names options requested_ranges successful_ranges]],partitionKeyColumns=[parent_id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[requested_ranges, exception_message, keyspace_name, successful_ranges, started_at, finished_at, options, exception_stacktrace, parent_id, columnfamily_names],droppedColumns={},triggers=[],indexes=[]], org.apache.cassandra.config.CFMetaData@5cff59c3[cfId=5582b59f-8e4e-35e1-b913-3acada51eb04,ksName=system_distributed,cfName=view_build_status,flags=[COMPOUND],params=TableParams{comment=Materialized View build status, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=3600000, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.UUIDType),partitionColumns=[[] | [status]],partitionKeyColumns=[keyspace_name, view_name],clusteringColumns=[host_id],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[view_name, status, keyspace_name, host_id],droppedColumns={},triggers=[],indexes=[]]], views=[], functions=[], types=[]} INFO [MigrationStage:1] 2025-09-18 05:23:13,551 ViewManager.java:137 - Not submitting build tasks for views in keyspace system_distributed as storage service is not initialized INFO [MigrationStage:1] 2025-09-18 05:23:13,558 ColumnFamilyStore.java:411 - Initializing system_distributed.parent_repair_history INFO [MigrationStage:1] 2025-09-18 05:23:13,576 ColumnFamilyStore.java:411 - Initializing system_distributed.repair_history INFO [MigrationStage:1] 2025-09-18 05:23:13,585 ColumnFamilyStore.java:411 - Initializing system_distributed.view_build_status INFO [main] 2025-09-18 05:23:13,613 StorageService.java:1446 - JOINING: Finish joining ring INFO [HANDSHAKE-/10.0.0.22] 2025-09-18 05:23:13,633 OutboundTcpConnection.java:561 - Handshaking version with /10.0.0.22 INFO [HANDSHAKE-/10.0.0.33] 2025-09-18 05:23:13,729 OutboundTcpConnection.java:561 - Handshaking version with /10.0.0.33 INFO [main] 2025-09-18 05:23:13,778 MigrationManager.java:331 - Create new Keyspace: KeyspaceMetadata{name=system_auth, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=1}}, tables=[org.apache.cassandra.config.CFMetaData@1a507b5c[cfId=5bc52802-de25-35ed-aeab-188eecebb090,ksName=system_auth,cfName=roles,flags=[COMPOUND],params=TableParams{comment=role definitions, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=7776000, default_time_to_live=0, memtable_flush_period_in_ms=3600000, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [can_login is_superuser salted_hash member_of]],partitionKeyColumns=[role],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UTF8Type,columnMetadata=[salted_hash, member_of, role, can_login, is_superuser],droppedColumns={},triggers=[],indexes=[]], org.apache.cassandra.config.CFMetaData@1552b2b0[cfId=0ecdaa87-f8fb-3e60-88d1-74fb36fe5c0d,ksName=system_auth,cfName=role_members,flags=[COMPOUND],params=TableParams{comment=role memberships lookup table, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=7776000, default_time_to_live=0, memtable_flush_period_in_ms=3600000, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | []],partitionKeyColumns=[role],clusteringColumns=[member],keyValidator=org.apache.cassandra.db.marshal.UTF8Type,columnMetadata=[role, member],droppedColumns={},triggers=[],indexes=[]], org.apache.cassandra.config.CFMetaData@19933c8e[cfId=3afbe79f-2194-31a7-add7-f5ab90d8ec9c,ksName=system_auth,cfName=role_permissions,flags=[COMPOUND],params=TableParams{comment=permissions granted to db roles, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=7776000, default_time_to_live=0, memtable_flush_period_in_ms=3600000, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [permissions]],partitionKeyColumns=[role],clusteringColumns=[resource],keyValidator=org.apache.cassandra.db.marshal.UTF8Type,columnMetadata=[role, resource, permissions],droppedColumns={},triggers=[],indexes=[]], org.apache.cassandra.config.CFMetaData@368f561[cfId=5f2fbdad-91f1-3946-bd25-d5da3a5c35ec,ksName=system_auth,cfName=resource_role_permissons_index,flags=[COMPOUND],params=TableParams{comment=index of db roles with permissions granted on a resource, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=7776000, default_time_to_live=0, memtable_flush_period_in_ms=3600000, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | []],partitionKeyColumns=[resource],clusteringColumns=[role],keyValidator=org.apache.cassandra.db.marshal.UTF8Type,columnMetadata=[resource, role],droppedColumns={},triggers=[],indexes=[]]], views=[], functions=[], types=[]} INFO [MigrationStage:1] 2025-09-18 05:23:13,984 ViewManager.java:137 - Not submitting build tasks for views in keyspace system_auth as storage service is not initialized INFO [MigrationStage:1] 2025-09-18 05:23:13,986 ColumnFamilyStore.java:411 - Initializing system_auth.resource_role_permissons_index INFO [MigrationStage:1] 2025-09-18 05:23:14,037 ColumnFamilyStore.java:411 - Initializing system_auth.role_members INFO [MigrationStage:1] 2025-09-18 05:23:14,042 ColumnFamilyStore.java:411 - Initializing system_auth.role_permissions INFO [MigrationStage:1] 2025-09-18 05:23:14,054 ColumnFamilyStore.java:411 - Initializing system_auth.roles INFO [main] 2025-09-18 05:23:14,069 Gossiper.java:1692 - Waiting for gossip to settle... INFO [GossipStage:1] 2025-09-18 05:23:14,755 Gossiper.java:1055 - Node /10.0.0.33 is now part of the cluster INFO [GossipStage:1] 2025-09-18 05:23:14,757 Gossiper.java:1055 - Node /10.0.0.22 is now part of the cluster WARN [MigrationStage:1] 2025-09-18 05:23:14,775 MigrationTask.java:67 - Can't send schema pull request: node /10.0.0.22 is down. INFO [GossipStage:1] 2025-09-18 05:23:14,775 TokenMetadata.java:479 - Updating topology for /10.0.0.22 INFO [GossipStage:1] 2025-09-18 05:23:14,775 TokenMetadata.java:479 - Updating topology for /10.0.0.22 WARN [MigrationStage:1] 2025-09-18 05:23:14,775 MigrationTask.java:67 - Can't send schema pull request: node /10.0.0.22 is down. INFO [RequestResponseStage-3] 2025-09-18 05:23:14,788 Gossiper.java:1019 - InetAddress /10.0.0.22 is now UP INFO [RequestResponseStage-6] 2025-09-18 05:23:14,802 Gossiper.java:1019 - InetAddress /10.0.0.33 is now UP WARN [GossipTasks:1] 2025-09-18 05:23:15,618 FailureDetector.java:288 - Not marking nodes down due to local pause of 5637406548 > 5000000000 INFO [main] 2025-09-18 05:23:22,070 Gossiper.java:1723 - No gossip backlog; proceeding INFO [main] 2025-09-18 05:23:22,402 NativeTransportService.java:70 - Netty using native Epoll event loop INFO [main] 2025-09-18 05:23:22,527 Server.java:155 - Using Netty Version: [netty-buffer=netty-buffer-4.1.39.Final.88c2a4c (repository: dirty), netty-codec=netty-codec-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-dns=netty-codec-dns-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-haproxy=netty-codec-haproxy-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-http=netty-codec-http-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-http2=netty-codec-http2-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-memcache=netty-codec-memcache-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-mqtt=netty-codec-mqtt-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-redis=netty-codec-redis-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-smtp=netty-codec-smtp-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-socks=netty-codec-socks-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-stomp=netty-codec-stomp-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-xml=netty-codec-xml-4.1.39.Final.88c2a4c (repository: dirty), netty-common=netty-common-4.1.39.Final.88c2a4c (repository: dirty), netty-handler=netty-handler-4.1.39.Final.88c2a4c (repository: dirty), netty-handler-proxy=netty-handler-proxy-4.1.39.Final.88c2a4c (repository: dirty), netty-resolver=netty-resolver-4.1.39.Final.88c2a4c (repository: dirty), netty-resolver-dns=netty-resolver-dns-4.1.39.Final.88c2a4c (repository: dirty), netty-tcnative=netty-tcnative-2.0.25.Final.c46c351, netty-transport=netty-transport-4.1.39.Final.88c2a4c (repository: dirty), netty-transport-native-epoll=netty-transport-native-epoll-4.1.39.Final.88c2a4c (repository: dirty), netty-transport-native-kqueue=netty-transport-native-kqueue-4.1.39.Final.88c2a4cab5 (repository: dirty), netty-transport-native-unix-common=netty-transport-native-unix-common-4.1.39.Final.88c2a4c (repository: dirty), netty-transport-rxtx=netty-transport-rxtx-4.1.39.Final.88c2a4c (repository: dirty), netty-transport-sctp=netty-transport-sctp-4.1.39.Final.88c2a4c (repository: dirty), netty-transport-udt=netty-transport-udt-4.1.39.Final.88c2a4c (repository: dirty)] INFO [main] 2025-09-18 05:23:22,529 Server.java:156 - Starting listening for CQL clients on /10.0.0.38:9041 (unencrypted)... INFO [main] 2025-09-18 05:23:22,608 ThriftServer.java:116 - Binding thrift service to /10.0.0.38:9161 INFO [Thread-2] 2025-09-18 05:23:22,614 ThriftServer.java:133 - Listening for thrift clients... INFO [OptionalTasks:1] 2025-09-18 05:23:24,264 CassandraRoleManager.java:356 - Created default superuser role 'cassandra' INFO [MigrationStage:1] 2025-09-18 05:23:30,135 ColumnFamilyStore.java:411 - Initializing reaper_db.schema_migration INFO [Native-Transport-Requests-1] 2025-09-18 05:23:30,833 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@1875b54f[cfId=9d261010-944f-11f0-9ace-ef813b7a18da,ksName=reaper_db,cfName=schema_migration_leader,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [leader leader_hostname took_lead_at]],partitionKeyColumns=[keyspace_name],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UTF8Type,columnMetadata=[leader, keyspace_name, took_lead_at, leader_hostname],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-09-18 05:23:31,063 ColumnFamilyStore.java:411 - Initializing reaper_db.schema_migration_leader INFO [HANDSHAKE-/10.0.0.38] 2025-09-18 05:23:32,586 OutboundTcpConnection.java:561 - Handshaking version with /10.0.0.38 WARN [Native-Transport-Requests-2] 2025-09-18 05:23:32,596 TimeFcts.java:99 - The function 'dateof' is deprecated. Use the function 'toTimestamp' instead. INFO [Native-Transport-Requests-1] 2025-09-18 05:23:32,637 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@64e75102[cfId=9e3954d0-944f-11f0-9ace-ef813b7a18da,ksName=reaper_db,cfName=running_reapers,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=180, default_time_to_live=180, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [last_heartbeat reaper_instance_host]],partitionKeyColumns=[reaper_instance_id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UUIDType,columnMetadata=[reaper_instance_id, last_heartbeat, reaper_instance_host],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-09-18 05:23:32,738 ColumnFamilyStore.java:411 - Initializing reaper_db.running_reapers INFO [MigrationStage:1] 2025-09-18 05:23:33,887 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_unit_v1 INFO [Native-Transport-Requests-1] 2025-09-18 05:23:34,748 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@769495d1[cfId=9f7b71c0-944f-11f0-9ace-ef813b7a18da,ksName=reaper_db,cfName=repair_schedule_by_cluster_and_keyspace,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[] | []],partitionKeyColumns=[cluster_name, keyspace_name],clusteringColumns=[repair_schedule_id],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[cluster_name, repair_schedule_id, keyspace_name],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-09-18 05:23:34,880 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_schedule_by_cluster_and_keyspace INFO [MigrationStage:1] 2025-09-18 05:23:35,819 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_run_by_cluster INFO [Native-Transport-Requests-1] 2025-09-18 05:23:36,746 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@5c50b561[cfId=a0ac2990-944f-11f0-9ace-ef813b7a18da,ksName=reaper_db,cfName=repair_schedule_v1,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [creation_time days_between intensity next_activation owner pause_time repair_parallelism repair_unit_id segment_count segment_count_per_node state run_history]],partitionKeyColumns=[id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[intensity, segment_count, days_between, id, state, run_history, creation_time, owner, repair_parallelism, segment_count_per_node, pause_time, repair_unit_id, next_activation],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-09-18 05:23:36,850 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_schedule_v1 INFO [MigrationStage:1] 2025-09-18 05:23:37,801 ColumnFamilyStore.java:411 - Initializing reaper_db.cluster INFO [Native-Transport-Requests-1] 2025-09-18 05:23:38,664 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@b2aeed1[cfId=a1d0fa80-944f-11f0-9ace-ef813b7a18da,ksName=reaper_db,cfName=snapshot,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'ALL'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [cause creation_time owner]],partitionKeyColumns=[cluster, snapshot_name],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[cluster, owner, cause, creation_time, snapshot_name],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-09-18 05:23:38,761 ColumnFamilyStore.java:411 - Initializing reaper_db.snapshot INFO [MigrationStage:1] 2025-09-18 05:23:39,894 ColumnFamilyStore.java:411 - Initializing reaper_db.node_metrics_v1 INFO [Native-Transport-Requests-1] 2025-09-18 05:23:40,784 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@1861d1e9[cfId=a3147700-944f-11f0-9ace-ef813b7a18da,ksName=reaper_db,cfName=repair_run,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : '5000'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[cause cluster_name creation_time end_time intensity last_event owner pause_time repair_parallelism repair_unit_id segment_count start_time state] | [coordinator_host end_token fail_count segment_end_time segment_start_time segment_state start_token token_ranges]],partitionKeyColumns=[id],clusteringColumns=[segment_id],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[intensity, segment_count, last_event, id, segment_end_time, state, cluster_name, end_time, end_token, start_token, segment_start_time, segment_state, cause, creation_time, start_time, coordinator_host, token_ranges, owner, repair_parallelism, segment_id, pause_time, repair_unit_id, fail_count],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-09-18 05:23:40,899 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_run INFO [MigrationStage:1] 2025-09-18 05:23:41,838 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_run_by_unit INFO [Native-Transport-Requests-1] 2025-09-18 05:23:42,760 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@7b203e78[cfId=a441fa80-944f-11f0-9ace-ef813b7a18da,ksName=reaper_db,cfName=leader,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=600, default_time_to_live=600, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [last_heartbeat reaper_instance_host reaper_instance_id]],partitionKeyColumns=[leader_id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[reaper_instance_id, last_heartbeat, reaper_instance_host, leader_id],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-09-18 05:23:42,871 ColumnFamilyStore.java:411 - Initializing reaper_db.leader INFO [Native-Transport-Requests-1] 2025-09-18 05:23:43,996 MigrationManager.java:454 - Update table 'reaper_db/cluster' From org.apache.cassandra.config.CFMetaData@75623166[cfId=a13b4a30-944f-11f0-8017-f51cf2f0029c,ksName=reaper_db,cfName=cluster,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'ALL'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [partitioner seed_hosts]],partitionKeyColumns=[name],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UTF8Type,columnMetadata=[partitioner, seed_hosts, name],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@bc0652d[cfId=a13b4a30-944f-11f0-8017-f51cf2f0029c,ksName=reaper_db,cfName=cluster,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'ALL'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [partitioner properties seed_hosts]],partitionKeyColumns=[name],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UTF8Type,columnMetadata=[properties, name, partitioner, seed_hosts],droppedColumns={},triggers=[],indexes=[]] INFO [STREAM-INIT-/10.0.0.33:47264] 2025-09-18 05:23:44,680 StreamResultFuture.java:116 - [Stream #a54ed6a0-944f-11f0-89ee-f38b18aca368 ID#0] Creating new streaming plan for Bootstrap INFO [STREAM-INIT-/10.0.0.33:47264] 2025-09-18 05:23:44,687 StreamResultFuture.java:123 - [Stream #a54ed6a0-944f-11f0-89ee-f38b18aca368, ID#0] Received streaming plan for Bootstrap INFO [STREAM-INIT-/10.0.0.33:47266] 2025-09-18 05:23:44,687 StreamResultFuture.java:123 - [Stream #a54ed6a0-944f-11f0-89ee-f38b18aca368, ID#0] Received streaming plan for Bootstrap INFO [STREAM-IN-/10.0.0.33:47266] 2025-09-18 05:23:44,839 StreamResultFuture.java:173 - [Stream #a54ed6a0-944f-11f0-89ee-f38b18aca368 ID#0] Prepare completed. Receiving 0 files(0.000KiB), sending 1 files(0.078KiB) INFO [STREAM-IN-/10.0.0.33:47266] 2025-09-18 05:23:44,903 StreamResultFuture.java:187 - [Stream #a54ed6a0-944f-11f0-89ee-f38b18aca368] Session with /10.0.0.33 is complete INFO [STREAM-IN-/10.0.0.33:47266] 2025-09-18 05:23:44,905 StreamResultFuture.java:219 - [Stream #a54ed6a0-944f-11f0-89ee-f38b18aca368] All sessions completed INFO [Native-Transport-Requests-2] 2025-09-18 05:23:45,278 MigrationManager.java:454 - Update table 'reaper_db/repair_run' From org.apache.cassandra.config.CFMetaData@6912b2a6[cfId=a3147700-944f-11f0-9ace-ef813b7a18da,ksName=reaper_db,cfName=repair_run,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : '5000'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[cause cluster_name creation_time end_time intensity last_event owner pause_time repair_parallelism repair_unit_id segment_count start_time state] | [coordinator_host end_token fail_count segment_end_time segment_start_time segment_state start_token token_ranges]],partitionKeyColumns=[id],clusteringColumns=[segment_id],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[intensity, segment_count, last_event, id, segment_end_time, state, cluster_name, end_time, end_token, start_token, segment_start_time, segment_state, cause, creation_time, start_time, coordinator_host, token_ranges, owner, repair_parallelism, segment_id, pause_time, repair_unit_id, fail_count],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@7727ce30[cfId=a3147700-944f-11f0-9ace-ef813b7a18da,ksName=reaper_db,cfName=repair_run,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : '5000'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[cause cluster_name creation_time end_time intensity last_event owner pause_time repair_parallelism repair_unit_id segment_count start_time state tables] | [coordinator_host end_token fail_count segment_end_time segment_start_time segment_state start_token token_ranges]],partitionKeyColumns=[id],clusteringColumns=[segment_id],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[intensity, segment_count, last_event, id, segment_end_time, state, cluster_name, end_time, end_token, start_token, segment_start_time, segment_state, cause, creation_time, start_time, coordinator_host, token_ranges, owner, repair_parallelism, tables, segment_id, pause_time, repair_unit_id, fail_count],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-4] 2025-09-18 05:23:45,877 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@24e3d29b[cfId=a61d9850-944f-11f0-9ace-ef813b7a18da,ksName=reaper_db,cfName=node_metrics_v2,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=300, default_time_to_live=864000, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4, unchecked_tombstone_compaction=true}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.UTF8Type, org.apache.cassandra.db.marshal.UTF8Type, org.apache.cassandra.db.marshal.UTF8Type, org.apache.cassandra.db.marshal.ReversedType(org.apache.cassandra.db.marshal.TimestampType), org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [value]],partitionKeyColumns=[cluster, metric_domain, metric_type, time_bucket],clusteringColumns=[host, metric_scope, metric_name, ts, metric_attribute],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[cluster, metric_domain, metric_attribute, time_bucket, ts, metric_type, metric_name, metric_scope, value, host],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-09-18 05:23:45,980 ColumnFamilyStore.java:411 - Initializing reaper_db.node_metrics_v2 INFO [MigrationStage:1] 2025-09-18 05:23:46,935 ColumnFamilyStore.java:411 - Initializing reaper_db.node_operations INFO [Native-Transport-Requests-1] 2025-09-18 05:23:48,728 MigrationManager.java:454 - Update table 'reaper_db/cluster' From org.apache.cassandra.config.CFMetaData@75623166[cfId=a13b4a30-944f-11f0-8017-f51cf2f0029c,ksName=reaper_db,cfName=cluster,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'ALL'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [partitioner properties state seed_hosts]],partitionKeyColumns=[name],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UTF8Type,columnMetadata=[properties, state, name, partitioner, seed_hosts],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@48bfddee[cfId=a13b4a30-944f-11f0-8017-f51cf2f0029c,ksName=reaper_db,cfName=cluster,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'ALL'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [last_contact partitioner properties state seed_hosts]],partitionKeyColumns=[name],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UTF8Type,columnMetadata=[properties, state, name, last_contact, partitioner, seed_hosts],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-1] 2025-09-18 05:23:49,775 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@71a960ad[cfId=a87061f0-944f-11f0-9ace-ef813b7a18da,ksName=reaper_db,cfName=diagnostic_event_subscription,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'ALL'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [cluster description export_file_logger export_http_endpoint export_sse events nodes]],partitionKeyColumns=[id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UUIDType,columnMetadata=[cluster, export_http_endpoint, events, id, export_sse, nodes, export_file_logger, description],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-09-18 05:23:49,901 ColumnFamilyStore.java:411 - Initializing reaper_db.diagnostic_event_subscription INFO [Native-Transport-Requests-2] 2025-09-18 05:23:50,873 MigrationManager.java:519 - Drop table 'reaper_db/node_metrics_v2' INFO [MigrationStage:1] 2025-09-18 05:23:51,879 ColumnFamilyStore.java:411 - Initializing reaper_db.node_metrics_v3 INFO [MigrationStage:1] 2025-09-18 05:23:53,025 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_run_by_cluster_v2 INFO [Native-Transport-Requests-1] 2025-09-18 05:23:54,928 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@7aec538c[cfId=ab82ab00-944f-11f0-9ace-ef813b7a18da,ksName=reaper_db,cfName=running_repairs,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=300, default_time_to_live=300, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [reaper_instance_host reaper_instance_id segment_id]],partitionKeyColumns=[repair_id],clusteringColumns=[node],keyValidator=org.apache.cassandra.db.marshal.UUIDType,columnMetadata=[reaper_instance_id, repair_id, node, segment_id, reaper_instance_host],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-09-18 05:23:55,027 ColumnFamilyStore.java:411 - Initializing reaper_db.running_repairs INFO [MigrationStage:1] 2025-09-18 05:23:57,314 ColumnFamilyStore.java:411 - Initializing reaper_db.percent_repaired_by_schedule INFO [Native-Transport-Requests-1] 2025-09-18 05:23:59,783 MigrationManager.java:454 - Update table 'reaper_db/repair_schedule_v1' From org.apache.cassandra.config.CFMetaData@253bb129[cfId=a0ac2990-944f-11f0-9ace-ef813b7a18da,ksName=reaper_db,cfName=repair_schedule_v1,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [creation_time days_between intensity next_activation owner pause_time repair_parallelism repair_unit_id segment_count segment_count_per_node state run_history]],partitionKeyColumns=[id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[intensity, segment_count, days_between, id, state, run_history, creation_time, owner, repair_parallelism, segment_count_per_node, pause_time, repair_unit_id, next_activation],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@5fa21ebd[cfId=a0ac2990-944f-11f0-9ace-ef813b7a18da,ksName=reaper_db,cfName=repair_schedule_v1,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [adaptive creation_time days_between intensity next_activation owner pause_time repair_parallelism repair_unit_id segment_count segment_count_per_node state run_history]],partitionKeyColumns=[id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[intensity, segment_count, days_between, id, state, run_history, creation_time, adaptive, owner, repair_parallelism, segment_count_per_node, pause_time, repair_unit_id, next_activation],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-1] 2025-09-18 05:24:02,753 MigrationManager.java:454 - Update table 'reaper_db/repair_schedule_v1' From org.apache.cassandra.config.CFMetaData@253bb129[cfId=a0ac2990-944f-11f0-9ace-ef813b7a18da,ksName=reaper_db,cfName=repair_schedule_v1,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [adaptive creation_time days_between intensity next_activation owner pause_time percent_unrepaired_threshold repair_parallelism repair_unit_id segment_count segment_count_per_node state run_history]],partitionKeyColumns=[id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[intensity, segment_count, days_between, percent_unrepaired_threshold, id, state, run_history, creation_time, adaptive, owner, repair_parallelism, segment_count_per_node, pause_time, repair_unit_id, next_activation],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@e65850d[cfId=a0ac2990-944f-11f0-9ace-ef813b7a18da,ksName=reaper_db,cfName=repair_schedule_v1,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [adaptive creation_time days_between intensity last_run next_activation owner pause_time percent_unrepaired_threshold repair_parallelism repair_unit_id segment_count segment_count_per_node state run_history]],partitionKeyColumns=[id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[intensity, segment_count, days_between, percent_unrepaired_threshold, id, last_run, state, run_history, creation_time, adaptive, owner, repair_parallelism, segment_count_per_node, pause_time, repair_unit_id, next_activation],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-1] 2025-09-18 05:24:06,795 MigrationManager.java:331 - Create new Keyspace: KeyspaceMetadata{name=config_db_uuid, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=3}}, tables=[], views=[], functions=[], types=[]} INFO [Native-Transport-Requests-8] 2025-09-18 05:24:06,824 MigrationManager.java:454 - Update table 'reaper_db/snapshot' From org.apache.cassandra.config.CFMetaData@7e29bee6[cfId=a1d0fa80-944f-11f0-9ace-ef813b7a18da,ksName=reaper_db,cfName=snapshot,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'ALL'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [cause creation_time owner]],partitionKeyColumns=[cluster, snapshot_name],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[cluster, owner, cause, creation_time, snapshot_name],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@2e5440e1[cfId=a1d0fa80-944f-11f0-9ace-ef813b7a18da,ksName=reaper_db,cfName=snapshot,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'ALL'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [cause creation_time owner]],partitionKeyColumns=[cluster, snapshot_name],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[cluster, owner, cause, creation_time, snapshot_name],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-6] 2025-09-18 05:24:06,824 MigrationManager.java:454 - Update table 'reaper_db/running_reapers' From org.apache.cassandra.config.CFMetaData@534fa7a7[cfId=9e3954d0-944f-11f0-9ace-ef813b7a18da,ksName=reaper_db,cfName=running_reapers,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=180, default_time_to_live=180, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [last_heartbeat reaper_instance_host]],partitionKeyColumns=[reaper_instance_id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UUIDType,columnMetadata=[reaper_instance_id, last_heartbeat, reaper_instance_host],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@556473d8[cfId=9e3954d0-944f-11f0-9ace-ef813b7a18da,ksName=reaper_db,cfName=running_reapers,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=180, default_time_to_live=180, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [last_heartbeat reaper_instance_host]],partitionKeyColumns=[reaper_instance_id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UUIDType,columnMetadata=[reaper_instance_id, last_heartbeat, reaper_instance_host],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-7] 2025-09-18 05:24:06,824 MigrationManager.java:454 - Update table 'reaper_db/percent_repaired_by_schedule' From org.apache.cassandra.config.CFMetaData@5c355d4d[cfId=ac9ddf00-944f-11f0-89ee-f38b18aca368,ksName=reaper_db,cfName=percent_repaired_by_schedule,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=300, default_time_to_live=3600, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [keyspace_name percent_repaired table_name ts]],partitionKeyColumns=[cluster_name, repair_schedule_id, time_bucket],clusteringColumns=[node],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UUIDType,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[time_bucket, node, ts, keyspace_name, percent_repaired, repair_schedule_id, table_name, cluster_name],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@3924ffcd[cfId=ac9ddf00-944f-11f0-89ee-f38b18aca368,ksName=reaper_db,cfName=percent_repaired_by_schedule,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=300, default_time_to_live=3600, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [keyspace_name percent_repaired table_name ts]],partitionKeyColumns=[cluster_name, repair_schedule_id, time_bucket],clusteringColumns=[node],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UUIDType,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[time_bucket, node, ts, keyspace_name, percent_repaired, repair_schedule_id, table_name, cluster_name],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-5] 2025-09-18 05:24:06,823 MigrationManager.java:454 - Update table 'reaper_db/repair_run_by_cluster_v2' From org.apache.cassandra.config.CFMetaData@7e53d82f[cfId=aa4cc310-944f-11f0-8017-f51cf2f0029c,ksName=reaper_db,cfName=repair_run_by_cluster_v2,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.ReversedType(org.apache.cassandra.db.marshal.TimeUUIDType)),partitionColumns=[[] | [repair_run_state]],partitionKeyColumns=[cluster_name],clusteringColumns=[id],keyValidator=org.apache.cassandra.db.marshal.UTF8Type,columnMetadata=[cluster_name, repair_run_state, id],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@7ae5e606[cfId=aa4cc310-944f-11f0-8017-f51cf2f0029c,ksName=reaper_db,cfName=repair_run_by_cluster_v2,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.ReversedType(org.apache.cassandra.db.marshal.TimeUUIDType)),partitionColumns=[[] | [repair_run_state]],partitionKeyColumns=[cluster_name],clusteringColumns=[id],keyValidator=org.apache.cassandra.db.marshal.UTF8Type,columnMetadata=[cluster_name, repair_run_state, id],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-4] 2025-09-18 05:24:06,824 MigrationManager.java:454 - Update table 'reaper_db/node_metrics_v1' From org.apache.cassandra.config.CFMetaData@2c39527[cfId=a27db540-944f-11f0-8017-f51cf2f0029c,ksName=reaper_db,cfName=node_metrics_v1,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=120, default_time_to_live=180, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4, unchecked_tombstone_compaction=true}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [active_anticompactions cluster datacenter has_repair_running pending_compactions requested]],partitionKeyColumns=[run_id, time_partition],clusteringColumns=[node],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UUIDType,org.apache.cassandra.db.marshal.LongType),columnMetadata=[cluster, node, has_repair_running, pending_compactions, active_anticompactions, time_partition, datacenter, requested, run_id],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@4fd58663[cfId=a27db540-944f-11f0-8017-f51cf2f0029c,ksName=reaper_db,cfName=node_metrics_v1,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=120, default_time_to_live=180, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4, unchecked_tombstone_compaction=true}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [active_anticompactions cluster datacenter has_repair_running pending_compactions requested]],partitionKeyColumns=[run_id, time_partition],clusteringColumns=[node],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UUIDType,org.apache.cassandra.db.marshal.LongType),columnMetadata=[cluster, node, has_repair_running, pending_compactions, active_anticompactions, time_partition, datacenter, requested, run_id],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-2] 2025-09-18 05:24:06,825 MigrationManager.java:454 - Update table 'reaper_db/leader' From org.apache.cassandra.config.CFMetaData@153aea89[cfId=a441fa80-944f-11f0-9ace-ef813b7a18da,ksName=reaper_db,cfName=leader,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=600, default_time_to_live=600, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [last_heartbeat reaper_instance_host reaper_instance_id]],partitionKeyColumns=[leader_id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[reaper_instance_id, last_heartbeat, reaper_instance_host, leader_id],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@1ed33036[cfId=a441fa80-944f-11f0-9ace-ef813b7a18da,ksName=reaper_db,cfName=leader,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=600, default_time_to_live=600, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [last_heartbeat reaper_instance_host reaper_instance_id]],partitionKeyColumns=[leader_id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[reaper_instance_id, last_heartbeat, reaper_instance_host, leader_id],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-3] 2025-09-18 05:24:06,822 MigrationManager.java:454 - Update table 'reaper_db/schema_migration_leader' From org.apache.cassandra.config.CFMetaData@724308a6[cfId=9d261010-944f-11f0-9ace-ef813b7a18da,ksName=reaper_db,cfName=schema_migration_leader,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [leader leader_hostname took_lead_at]],partitionKeyColumns=[keyspace_name],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UTF8Type,columnMetadata=[leader, keyspace_name, took_lead_at, leader_hostname],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@2787988f[cfId=9d261010-944f-11f0-9ace-ef813b7a18da,ksName=reaper_db,cfName=schema_migration_leader,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [leader leader_hostname took_lead_at]],partitionKeyColumns=[keyspace_name],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UTF8Type,columnMetadata=[leader, keyspace_name, took_lead_at, leader_hostname],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-09-18 05:24:11,772 ColumnFamilyStore.java:411 - Initializing svc_monitor_keyspace.service_instance_table INFO [Native-Transport-Requests-1] 2025-09-18 05:24:12,876 MigrationManager.java:454 - Update table 'reaper_db/node_metrics_v3' From org.apache.cassandra.config.CFMetaData@488293c[cfId=a99f1df0-944f-11f0-8017-f51cf2f0029c,ksName=reaper_db,cfName=node_metrics_v3,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=300, default_time_to_live=3600, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4, unchecked_tombstone_compaction=true}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.ReversedType(org.apache.cassandra.db.marshal.TimestampType), org.apache.cassandra.db.marshal.UTF8Type, org.apache.cassandra.db.marshal.UTF8Type, org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [value]],partitionKeyColumns=[cluster, metric_domain, metric_type, time_bucket, host],clusteringColumns=[ts, metric_scope, metric_name, metric_attribute],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[cluster, metric_domain, metric_attribute, time_bucket, ts, metric_type, metric_name, metric_scope, value, host],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@d4c35ea[cfId=a99f1df0-944f-11f0-8017-f51cf2f0029c,ksName=reaper_db,cfName=node_metrics_v3,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=300, default_time_to_live=3600, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.TimeWindowCompactionStrategy, options={min_threshold=4, max_threshold=32, compaction_window_size=10, compaction_window_unit=MINUTES, unchecked_tombstone_compaction=true}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.ReversedType(org.apache.cassandra.db.marshal.TimestampType), org.apache.cassandra.db.marshal.UTF8Type, org.apache.cassandra.db.marshal.UTF8Type, org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [value]],partitionKeyColumns=[cluster, metric_domain, metric_type, time_bucket, host],clusteringColumns=[ts, metric_scope, metric_name, metric_attribute],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[cluster, metric_domain, metric_attribute, time_bucket, ts, metric_type, metric_name, metric_scope, value, host],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-2] 2025-09-18 05:24:12,985 MigrationManager.java:331 - Create new Keyspace: KeyspaceMetadata{name=useragent, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=3}}, tables=[], views=[], functions=[], types=[]} INFO [Native-Transport-Requests-1] 2025-09-18 05:24:15,436 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@1dedda64[cfId=b7bbf0c0-944f-11f0-9ace-ef813b7a18da,ksName=config_db_uuid,cfName=obj_uuid_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-09-18 05:24:15,543 ColumnFamilyStore.java:411 - Initializing config_db_uuid.obj_uuid_table INFO [Native-Transport-Requests-4] 2025-09-18 05:24:16,805 MigrationManager.java:427 - Update Keyspace 'useragent' From KeyspaceMetadata{name=useragent, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=3}}, tables=[], views=[], functions=[], types=[]} To KeyspaceMetadata{name=useragent, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=3}}, tables=[], views=[], functions=[], types=[]} INFO [Native-Transport-Requests-1] 2025-09-18 05:24:17,679 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@374861b8[cfId=b91231f0-944f-11f0-9ace-ef813b7a18da,ksName=svc_monitor_keyspace,cfName=pool_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-09-18 05:24:17,804 ColumnFamilyStore.java:411 - Initializing svc_monitor_keyspace.pool_table INFO [Native-Transport-Requests-1] 2025-09-18 05:24:20,492 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@5ccc6f61[cfId=babf45b0-944f-11f0-9ace-ef813b7a18da,ksName=config_db_uuid,cfName=obj_fq_name_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-09-18 05:24:20,603 ColumnFamilyStore.java:411 - Initializing config_db_uuid.obj_fq_name_table INFO [MigrationStage:1] 2025-09-18 05:24:23,611 ColumnFamilyStore.java:411 - Initializing svc_monitor_keyspace.loadbalancer_table INFO [Native-Transport-Requests-1] 2025-09-18 05:24:23,849 MigrationManager.java:454 - Update table 'config_db_uuid/obj_fq_name_table' From org.apache.cassandra.config.CFMetaData@55f47a8b[cfId=babf45b0-944f-11f0-9ace-ef813b7a18da,ksName=config_db_uuid,cfName=obj_fq_name_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@49d63d95[cfId=babf45b0-944f-11f0-9ace-ef813b7a18da,ksName=config_db_uuid,cfName=obj_fq_name_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-09-18 05:24:24,910 ColumnFamilyStore.java:411 - Initializing config_db_uuid.obj_shared_table INFO [Native-Transport-Requests-1] 2025-09-18 05:24:25,738 MigrationManager.java:454 - Update table 'svc_monitor_keyspace/loadbalancer_table' From org.apache.cassandra.config.CFMetaData@1077df6f[cfId=bc8a41b0-944f-11f0-8017-f51cf2f0029c,ksName=svc_monitor_keyspace,cfName=loadbalancer_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@540d4e00[cfId=bc8a41b0-944f-11f0-8017-f51cf2f0029c,ksName=svc_monitor_keyspace,cfName=loadbalancer_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-2] 2025-09-18 05:24:26,817 MigrationManager.java:454 - Update table 'config_db_uuid/obj_shared_table' From org.apache.cassandra.config.CFMetaData@3308f1f3[cfId=bd4b20b0-944f-11f0-8017-f51cf2f0029c,ksName=config_db_uuid,cfName=obj_shared_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@15eb5e9e[cfId=bd4b20b0-944f-11f0-8017-f51cf2f0029c,ksName=config_db_uuid,cfName=obj_shared_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-09-18 05:24:29,267 ColumnFamilyStore.java:411 - Initializing svc_monitor_keyspace.healthmonitor_table INFO [MigrationStage:1] 2025-09-18 05:24:30,929 ColumnFamilyStore.java:411 - Initializing useragent.useragent_keyval_table INFO [Native-Transport-Requests-3] 2025-09-18 05:24:35,467 MigrationManager.java:454 - Update table 'config_db_uuid/obj_fq_name_table' From org.apache.cassandra.config.CFMetaData@55f47a8b[cfId=babf45b0-944f-11f0-9ace-ef813b7a18da,ksName=config_db_uuid,cfName=obj_fq_name_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@dd0319e[cfId=babf45b0-944f-11f0-9ace-ef813b7a18da,ksName=config_db_uuid,cfName=obj_fq_name_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-2] 2025-09-18 05:24:39,520 MigrationManager.java:427 - Update Keyspace 'to_bgp_keyspace' From KeyspaceMetadata{name=to_bgp_keyspace, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=3}}, tables=[], views=[], functions=[], types=[]} To KeyspaceMetadata{name=to_bgp_keyspace, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=3}}, tables=[], views=[], functions=[], types=[]} INFO [Native-Transport-Requests-1] 2025-09-18 05:24:40,875 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@53c0dc9f[cfId=c6e59fb0-944f-11f0-9ace-ef813b7a18da,ksName=to_bgp_keyspace,cfName=route_target_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-09-18 05:24:40,978 ColumnFamilyStore.java:411 - Initializing to_bgp_keyspace.route_target_table INFO [MigrationStage:1] 2025-09-18 05:24:42,916 ColumnFamilyStore.java:411 - Initializing to_bgp_keyspace.service_chain_ip_address_table INFO [Native-Transport-Requests-1] 2025-09-18 05:24:43,811 MigrationManager.java:454 - Update table 'to_bgp_keyspace/service_chain_ip_address_table' From org.apache.cassandra.config.CFMetaData@4b0f1a88[cfId=c80feee0-944f-11f0-89ee-f38b18aca368,ksName=to_bgp_keyspace,cfName=service_chain_ip_address_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@3580411a[cfId=c80feee0-944f-11f0-89ee-f38b18aca368,ksName=to_bgp_keyspace,cfName=service_chain_ip_address_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-09-18 05:24:44,805 ColumnFamilyStore.java:411 - Initializing to_bgp_keyspace.service_chain_table INFO [Native-Transport-Requests-1] 2025-09-18 05:24:47,489 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@14b7c274[cfId=cad6d710-944f-11f0-9ace-ef813b7a18da,ksName=to_bgp_keyspace,cfName=service_chain_uuid_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-09-18 05:24:47,594 ColumnFamilyStore.java:411 - Initializing to_bgp_keyspace.service_chain_uuid_table INFO [Native-Transport-Requests-1] 2025-09-18 05:25:56,885 MigrationManager.java:427 - Update Keyspace 'dm_keyspace' From KeyspaceMetadata{name=dm_keyspace, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=3}}, tables=[], views=[], functions=[], types=[]} To KeyspaceMetadata{name=dm_keyspace, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=3}}, tables=[], views=[], functions=[], types=[]} INFO [Native-Transport-Requests-1] 2025-09-18 05:25:57,842 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@581650f8[cfId=f4c5db20-944f-11f0-9ace-ef813b7a18da,ksName=dm_keyspace,cfName=dm_pr_vn_ip_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-09-18 05:25:57,939 ColumnFamilyStore.java:411 - Initializing dm_keyspace.dm_pr_vn_ip_table INFO [MigrationStage:1] 2025-09-18 05:26:00,038 ColumnFamilyStore.java:411 - Initializing dm_keyspace.dm_pr_asn_table INFO [Native-Transport-Requests-1] 2025-09-18 05:26:00,859 MigrationManager.java:454 - Update table 'dm_keyspace/dm_pr_asn_table' From org.apache.cassandra.config.CFMetaData@54a155b4[cfId=f601b680-944f-11f0-89ee-f38b18aca368,ksName=dm_keyspace,cfName=dm_pr_asn_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@479c4f23[cfId=f601b680-944f-11f0-89ee-f38b18aca368,ksName=dm_keyspace,cfName=dm_pr_asn_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-09-18 05:26:02,083 ColumnFamilyStore.java:411 - Initializing dm_keyspace.dm_ni_ipv6_ll_table INFO [Native-Transport-Requests-1] 2025-09-18 05:26:04,474 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@51b799f8[cfId=f8b9d1a0-944f-11f0-9ace-ef813b7a18da,ksName=dm_keyspace,cfName=dm_pnf_resource_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@97d350d4, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-09-18 05:26:04,577 ColumnFamilyStore.java:411 - Initializing dm_keyspace.dm_pnf_resource_table INFO [HANDSHAKE-/10.0.0.22] 2025-09-18 05:31:18,958 OutboundTcpConnection.java:561 - Handshaking version with /10.0.0.22 INFO [HANDSHAKE-/10.0.0.33] 2025-09-18 05:31:19,173 OutboundTcpConnection.java:561 - Handshaking version with /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:19,435 Validator.java:281 - [repair #b45f89e0-9450-11f0-8017-f51cf2f0029c] Sending completed merkle tree to /10.0.0.22 for reaper_db.schema_migration_leader INFO [AntiEntropyStage:1] 2025-09-18 05:31:19,467 Validator.java:281 - [repair #b45f89e0-9450-11f0-8017-f51cf2f0029c] Sending completed merkle tree to /10.0.0.22 for reaper_db.repair_run_by_cluster INFO [AntiEntropyStage:1] 2025-09-18 05:31:19,545 Validator.java:281 - [repair #b45f89e0-9450-11f0-8017-f51cf2f0029c] Sending completed merkle tree to /10.0.0.22 for reaper_db.repair_run_by_unit INFO [AntiEntropyStage:1] 2025-09-18 05:31:19,687 Validator.java:281 - [repair #b45f89e0-9450-11f0-8017-f51cf2f0029c] Sending completed merkle tree to /10.0.0.22 for reaper_db.repair_schedule_v1 INFO [AntiEntropyStage:1] 2025-09-18 05:31:19,703 Validator.java:281 - [repair #b45f89e0-9450-11f0-8017-f51cf2f0029c] Sending completed merkle tree to /10.0.0.22 for reaper_db.leader INFO [AntiEntropyStage:1] 2025-09-18 05:31:19,767 Validator.java:281 - [repair #b45f89e0-9450-11f0-8017-f51cf2f0029c] Sending completed merkle tree to /10.0.0.22 for reaper_db.cluster INFO [AntiEntropyStage:1] 2025-09-18 05:31:19,795 Validator.java:281 - [repair #b45f89e0-9450-11f0-8017-f51cf2f0029c] Sending completed merkle tree to /10.0.0.22 for reaper_db.percent_repaired_by_schedule INFO [AntiEntropyStage:1] 2025-09-18 05:31:19,813 Validator.java:281 - [repair #b45f89e0-9450-11f0-8017-f51cf2f0029c] Sending completed merkle tree to /10.0.0.22 for reaper_db.schema_migration INFO [AntiEntropyStage:1] 2025-09-18 05:31:19,894 Validator.java:281 - [repair #b45f89e0-9450-11f0-8017-f51cf2f0029c] Sending completed merkle tree to /10.0.0.22 for reaper_db.running_repairs INFO [AntiEntropyStage:1] 2025-09-18 05:31:19,973 Validator.java:281 - [repair #b45f89e0-9450-11f0-8017-f51cf2f0029c] Sending completed merkle tree to /10.0.0.22 for reaper_db.repair_unit_v1 INFO [AntiEntropyStage:1] 2025-09-18 05:31:20,034 Validator.java:281 - [repair #b45f89e0-9450-11f0-8017-f51cf2f0029c] Sending completed merkle tree to /10.0.0.22 for reaper_db.repair_schedule_by_cluster_and_keyspace INFO [AntiEntropyStage:1] 2025-09-18 05:31:20,043 Validator.java:281 - [repair #b45f89e0-9450-11f0-8017-f51cf2f0029c] Sending completed merkle tree to /10.0.0.22 for reaper_db.diagnostic_event_subscription INFO [AntiEntropyStage:1] 2025-09-18 05:31:20,110 Validator.java:281 - [repair #b45f89e0-9450-11f0-8017-f51cf2f0029c] Sending completed merkle tree to /10.0.0.22 for reaper_db.repair_run INFO [AntiEntropyStage:1] 2025-09-18 05:31:20,163 Validator.java:281 - [repair #b45f89e0-9450-11f0-8017-f51cf2f0029c] Sending completed merkle tree to /10.0.0.22 for reaper_db.repair_run_by_cluster_v2 INFO [AntiEntropyStage:1] 2025-09-18 05:31:20,231 Validator.java:281 - [repair #b45f89e0-9450-11f0-8017-f51cf2f0029c] Sending completed merkle tree to /10.0.0.22 for reaper_db.running_reapers INFO [AntiEntropyStage:1] 2025-09-18 05:31:20,242 Validator.java:281 - [repair #b45f89e0-9450-11f0-8017-f51cf2f0029c] Sending completed merkle tree to /10.0.0.22 for reaper_db.snapshot INFO [AntiEntropyStage:1] 2025-09-18 05:31:20,249 ActiveRepairService.java:452 - [repair #b4524370-9450-11f0-8017-f51cf2f0029c] Not a global repair, will not do anticompaction INFO [AntiEntropyStage:1] 2025-09-18 05:31:28,332 Validator.java:281 - [repair #b9913080-9450-11f0-89ee-f38b18aca368] Sending completed merkle tree to /10.0.0.33 for reaper_db.repair_run_by_cluster INFO [AntiEntropyStage:1] 2025-09-18 05:31:28,355 Validator.java:281 - [repair #b9913080-9450-11f0-89ee-f38b18aca368] Sending completed merkle tree to /10.0.0.33 for reaper_db.snapshot INFO [AntiEntropyStage:1] 2025-09-18 05:31:28,403 Validator.java:281 - [repair #b9913080-9450-11f0-89ee-f38b18aca368] Sending completed merkle tree to /10.0.0.33 for reaper_db.repair_schedule_by_cluster_and_keyspace INFO [AntiEntropyStage:1] 2025-09-18 05:31:28,499 Validator.java:281 - [repair #b9913080-9450-11f0-89ee-f38b18aca368] Sending completed merkle tree to /10.0.0.33 for reaper_db.repair_run INFO [AntiEntropyStage:1] 2025-09-18 05:31:28,527 Validator.java:281 - [repair #b9913080-9450-11f0-89ee-f38b18aca368] Sending completed merkle tree to /10.0.0.33 for reaper_db.repair_run_by_cluster_v2 INFO [AntiEntropyStage:1] 2025-09-18 05:31:28,614 Validator.java:281 - [repair #b9913080-9450-11f0-89ee-f38b18aca368] Sending completed merkle tree to /10.0.0.33 for reaper_db.running_repairs INFO [AntiEntropyStage:1] 2025-09-18 05:31:28,671 Validator.java:281 - [repair #b9913080-9450-11f0-89ee-f38b18aca368] Sending completed merkle tree to /10.0.0.33 for reaper_db.schema_migration INFO [AntiEntropyStage:1] 2025-09-18 05:31:28,694 Validator.java:281 - [repair #b9913080-9450-11f0-89ee-f38b18aca368] Sending completed merkle tree to /10.0.0.33 for reaper_db.diagnostic_event_subscription INFO [AntiEntropyStage:1] 2025-09-18 05:31:28,704 Validator.java:281 - [repair #b9913080-9450-11f0-89ee-f38b18aca368] Sending completed merkle tree to /10.0.0.33 for reaper_db.schema_migration_leader INFO [AntiEntropyStage:1] 2025-09-18 05:31:28,718 Validator.java:281 - [repair #b9913080-9450-11f0-89ee-f38b18aca368] Sending completed merkle tree to /10.0.0.33 for reaper_db.leader INFO [AntiEntropyStage:1] 2025-09-18 05:31:28,760 Validator.java:281 - [repair #b9913080-9450-11f0-89ee-f38b18aca368] Sending completed merkle tree to /10.0.0.33 for reaper_db.repair_unit_v1 INFO [AntiEntropyStage:1] 2025-09-18 05:31:28,798 Validator.java:281 - [repair #b9913080-9450-11f0-89ee-f38b18aca368] Sending completed merkle tree to /10.0.0.33 for reaper_db.running_reapers INFO [AntiEntropyStage:1] 2025-09-18 05:31:28,808 Validator.java:281 - [repair #b9913080-9450-11f0-89ee-f38b18aca368] Sending completed merkle tree to /10.0.0.33 for reaper_db.percent_repaired_by_schedule INFO [AntiEntropyStage:1] 2025-09-18 05:31:28,841 Validator.java:281 - [repair #b9913080-9450-11f0-89ee-f38b18aca368] Sending completed merkle tree to /10.0.0.33 for reaper_db.repair_run_by_unit INFO [AntiEntropyStage:1] 2025-09-18 05:31:28,859 Validator.java:281 - [repair #b9913080-9450-11f0-89ee-f38b18aca368] Sending completed merkle tree to /10.0.0.33 for reaper_db.repair_schedule_v1 INFO [AntiEntropyStage:1] 2025-09-18 05:31:28,879 Validator.java:281 - [repair #b9913080-9450-11f0-89ee-f38b18aca368] Sending completed merkle tree to /10.0.0.33 for reaper_db.cluster INFO [AntiEntropyStage:1] 2025-09-18 05:31:28,897 ActiveRepairService.java:452 - [repair #b9815200-9450-11f0-89ee-f38b18aca368] Not a global repair, will not do anticompaction INFO [Repair-Task-2] 2025-09-18 05:31:29,322 RepairRunnable.java:139 - Starting repair command #1 (ba59b0a0-9450-11f0-9ace-ef813b7a18da), repairing keyspace reaper_db with repair options (parallelism: dc_parallel, primary range: false, incremental: false, job threads: 1, ColumnFamilies: [cluster, leader, diagnostic_event_subscription, schema_migration_leader, running_repairs, repair_run, repair_unit_v1, repair_run_by_cluster_v2, repair_schedule_by_cluster_and_keyspace, repair_run_by_cluster, running_reapers, repair_schedule_v1, percent_repaired_by_schedule, repair_run_by_unit, schema_migration, snapshot], dataCenters: [], hosts: [], # of ranges: 2, pull repair: false) INFO [Repair-Task-2] 2025-09-18 05:31:29,373 RepairSession.java:228 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] new session: will sync /10.0.0.38, /10.0.0.33, /10.0.0.22 on range [(-9083771317662873172,-9064397874770026951], (5482710683617896693,5552469124657401434]] for reaper_db.[repair_run, snapshot, leader, repair_unit_v1, repair_schedule_by_cluster_and_keyspace, repair_schedule_v1, schema_migration_leader, repair_run_by_cluster_v2, schema_migration, running_reapers, repair_run_by_unit, cluster, running_repairs, percent_repaired_by_schedule, repair_run_by_cluster, diagnostic_event_subscription] INFO [RepairJobTask:2] 2025-09-18 05:31:29,467 RepairJob.java:234 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for repair_run (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:2] 2025-09-18 05:31:29,468 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,473 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,473 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,481 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,481 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,483 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run from /10.0.0.38 INFO [RepairJobTask:1] 2025-09-18 05:31:29,487 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for repair_run INFO [RepairJobTask:4] 2025-09-18 05:31:29,488 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for repair_run INFO [RepairJobTask:5] 2025-09-18 05:31:29,489 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for repair_run INFO [RepairJobTask:4] 2025-09-18 05:31:29,489 RepairJob.java:143 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] repair_run is fully synced INFO [RepairJobTask:4] 2025-09-18 05:31:29,492 RepairJob.java:234 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for snapshot (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:4] 2025-09-18 05:31:29,493 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,496 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for snapshot from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,496 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,498 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for snapshot from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,499 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,501 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for snapshot from /10.0.0.38 INFO [RepairJobTask:2] 2025-09-18 05:31:29,502 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for snapshot INFO [RepairJobTask:5] 2025-09-18 05:31:29,502 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for snapshot INFO [RepairJobTask:4] 2025-09-18 05:31:29,504 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for snapshot INFO [RepairJobTask:5] 2025-09-18 05:31:29,505 RepairJob.java:143 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] snapshot is fully synced INFO [RepairJobTask:2] 2025-09-18 05:31:29,505 RepairJob.java:234 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for leader (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:2] 2025-09-18 05:31:29,505 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,508 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for leader from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,508 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,513 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for leader from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,513 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,514 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for leader from /10.0.0.38 INFO [RepairJobTask:1] 2025-09-18 05:31:29,517 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for leader INFO [RepairJobTask:3] 2025-09-18 05:31:29,517 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for leader INFO [RepairJobTask:2] 2025-09-18 05:31:29,517 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for leader INFO [RepairJobTask:3] 2025-09-18 05:31:29,517 RepairJob.java:143 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] leader is fully synced INFO [RepairJobTask:3] 2025-09-18 05:31:29,523 RepairJob.java:234 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for repair_unit_v1 (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:3] 2025-09-18 05:31:29,523 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,531 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_unit_v1 from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,531 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,544 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_unit_v1 from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,544 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,547 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_unit_v1 from /10.0.0.38 INFO [RepairJobTask:5] 2025-09-18 05:31:29,548 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for repair_unit_v1 INFO [RepairJobTask:5] 2025-09-18 05:31:29,552 RepairJob.java:234 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for repair_schedule_by_cluster_and_keyspace (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:5] 2025-09-18 05:31:29,552 RepairJob.java:257 - Validating /10.0.0.33 INFO [RepairJobTask:1] 2025-09-18 05:31:29,552 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for repair_unit_v1 INFO [RepairJobTask:6] 2025-09-18 05:31:29,553 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for repair_unit_v1 INFO [RepairJobTask:1] 2025-09-18 05:31:29,553 RepairJob.java:143 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] repair_unit_v1 is fully synced INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,557 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,557 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,560 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,561 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,563 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.38 INFO [RepairJobTask:6] 2025-09-18 05:31:29,563 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:5] 2025-09-18 05:31:29,563 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:2] 2025-09-18 05:31:29,563 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:5] 2025-09-18 05:31:29,564 RepairJob.java:143 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] repair_schedule_by_cluster_and_keyspace is fully synced INFO [RepairJobTask:5] 2025-09-18 05:31:29,573 RepairJob.java:234 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for repair_schedule_v1 (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:5] 2025-09-18 05:31:29,573 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,578 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_schedule_v1 from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,580 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,584 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_schedule_v1 from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,584 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,588 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_schedule_v1 from /10.0.0.38 INFO [RepairJobTask:1] 2025-09-18 05:31:29,589 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for repair_schedule_v1 INFO [RepairJobTask:6] 2025-09-18 05:31:29,589 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for repair_schedule_v1 INFO [RepairJobTask:2] 2025-09-18 05:31:29,589 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for repair_schedule_v1 INFO [RepairJobTask:6] 2025-09-18 05:31:29,589 RepairJob.java:143 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] repair_schedule_v1 is fully synced INFO [RepairJobTask:6] 2025-09-18 05:31:29,594 RepairJob.java:234 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for schema_migration_leader (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:6] 2025-09-18 05:31:29,594 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,602 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for schema_migration_leader from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,602 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,606 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for schema_migration_leader from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,606 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,607 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for schema_migration_leader from /10.0.0.38 INFO [RepairJobTask:1] 2025-09-18 05:31:29,607 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for schema_migration_leader INFO [RepairJobTask:5] 2025-09-18 05:31:29,607 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for schema_migration_leader INFO [RepairJobTask:6] 2025-09-18 05:31:29,607 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for schema_migration_leader INFO [RepairJobTask:4] 2025-09-18 05:31:29,608 RepairJob.java:143 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] schema_migration_leader is fully synced INFO [RepairJobTask:4] 2025-09-18 05:31:29,612 RepairJob.java:234 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for repair_run_by_cluster_v2 (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:4] 2025-09-18 05:31:29,612 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,618 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,618 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,625 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,625 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,626 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.38 INFO [RepairJobTask:2] 2025-09-18 05:31:29,626 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:4] 2025-09-18 05:31:29,626 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:5] 2025-09-18 05:31:29,626 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:4] 2025-09-18 05:31:29,626 RepairJob.java:143 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] repair_run_by_cluster_v2 is fully synced INFO [RepairJobTask:4] 2025-09-18 05:31:29,630 RepairJob.java:234 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for schema_migration (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:4] 2025-09-18 05:31:29,631 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,636 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for schema_migration from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,637 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,640 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for schema_migration from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,640 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,642 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for schema_migration from /10.0.0.38 INFO [RepairJobTask:5] 2025-09-18 05:31:29,642 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for schema_migration INFO [RepairJobTask:4] 2025-09-18 05:31:29,642 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for schema_migration INFO [RepairJobTask:6] 2025-09-18 05:31:29,642 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for schema_migration INFO [RepairJobTask:4] 2025-09-18 05:31:29,642 RepairJob.java:143 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] schema_migration is fully synced INFO [RepairJobTask:4] 2025-09-18 05:31:29,656 RepairJob.java:234 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for running_reapers (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:4] 2025-09-18 05:31:29,656 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,660 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for running_reapers from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,660 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,665 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for running_reapers from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,665 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,671 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for running_reapers from /10.0.0.38 INFO [RepairJobTask:3] 2025-09-18 05:31:29,671 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for running_reapers INFO [RepairJobTask:1] 2025-09-18 05:31:29,671 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for running_reapers INFO [RepairJobTask:4] 2025-09-18 05:31:29,671 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for running_reapers INFO [RepairJobTask:1] 2025-09-18 05:31:29,671 RepairJob.java:143 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] running_reapers is fully synced INFO [RepairJobTask:4] 2025-09-18 05:31:29,677 RepairJob.java:234 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for repair_run_by_unit (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:4] 2025-09-18 05:31:29,677 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,680 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run_by_unit from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,680 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,684 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run_by_unit from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,684 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,686 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run_by_unit from /10.0.0.38 INFO [RepairJobTask:1] 2025-09-18 05:31:29,686 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for repair_run_by_unit INFO [RepairJobTask:4] 2025-09-18 05:31:29,686 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for repair_run_by_unit INFO [RepairJobTask:6] 2025-09-18 05:31:29,687 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for repair_run_by_unit INFO [RepairJobTask:4] 2025-09-18 05:31:29,687 RepairJob.java:143 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] repair_run_by_unit is fully synced INFO [RepairJobTask:4] 2025-09-18 05:31:29,696 RepairJob.java:234 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for cluster (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:4] 2025-09-18 05:31:29,696 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,703 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for cluster from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,703 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,705 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for cluster from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,705 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,707 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for cluster from /10.0.0.38 INFO [RepairJobTask:1] 2025-09-18 05:31:29,708 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for cluster INFO [RepairJobTask:6] 2025-09-18 05:31:29,708 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for cluster INFO [RepairJobTask:3] 2025-09-18 05:31:29,708 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for cluster INFO [RepairJobTask:6] 2025-09-18 05:31:29,708 RepairJob.java:143 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] cluster is fully synced INFO [RepairJobTask:6] 2025-09-18 05:31:29,755 RepairJob.java:234 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for running_repairs (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:6] 2025-09-18 05:31:29,755 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,761 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for running_repairs from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,762 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,763 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for running_repairs from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,763 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,765 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for running_repairs from /10.0.0.38 INFO [RepairJobTask:5] 2025-09-18 05:31:29,766 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for running_repairs INFO [RepairJobTask:6] 2025-09-18 05:31:29,766 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for running_repairs INFO [RepairJobTask:4] 2025-09-18 05:31:29,766 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for running_repairs INFO [RepairJobTask:6] 2025-09-18 05:31:29,766 RepairJob.java:143 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] running_repairs is fully synced INFO [RepairJobTask:7] 2025-09-18 05:31:29,770 RepairJob.java:234 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for percent_repaired_by_schedule (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:7] 2025-09-18 05:31:29,770 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,776 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for percent_repaired_by_schedule from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,776 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,778 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for percent_repaired_by_schedule from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,778 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,781 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for percent_repaired_by_schedule from /10.0.0.38 INFO [RepairJobTask:5] 2025-09-18 05:31:29,781 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:2] 2025-09-18 05:31:29,782 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:4] 2025-09-18 05:31:29,782 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:2] 2025-09-18 05:31:29,782 RepairJob.java:143 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] percent_repaired_by_schedule is fully synced INFO [RepairJobTask:2] 2025-09-18 05:31:29,796 RepairJob.java:234 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for repair_run_by_cluster (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:2] 2025-09-18 05:31:29,797 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,799 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run_by_cluster from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,799 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,802 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run_by_cluster from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,802 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,805 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run_by_cluster from /10.0.0.38 INFO [RepairJobTask:5] 2025-09-18 05:31:29,805 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for repair_run_by_cluster INFO [RepairJobTask:2] 2025-09-18 05:31:29,805 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for repair_run_by_cluster INFO [RepairJobTask:2] 2025-09-18 05:31:29,810 RepairJob.java:234 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for diagnostic_event_subscription (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:2] 2025-09-18 05:31:29,810 RepairJob.java:257 - Validating /10.0.0.33 INFO [RepairJobTask:7] 2025-09-18 05:31:29,812 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for repair_run_by_cluster INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,812 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for diagnostic_event_subscription from /10.0.0.33 INFO [RepairJobTask:5] 2025-09-18 05:31:29,812 RepairJob.java:143 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] repair_run_by_cluster is fully synced INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,813 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,815 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for diagnostic_event_subscription from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,815 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:29,817 RepairSession.java:180 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for diagnostic_event_subscription from /10.0.0.38 INFO [RepairJobTask:4] 2025-09-18 05:31:29,818 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for diagnostic_event_subscription INFO [RepairJobTask:2] 2025-09-18 05:31:29,818 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for diagnostic_event_subscription INFO [RepairJobTask:7] 2025-09-18 05:31:29,818 SyncTask.java:66 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for diagnostic_event_subscription INFO [RepairJobTask:2] 2025-09-18 05:31:29,818 RepairJob.java:143 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] diagnostic_event_subscription is fully synced INFO [RepairJobTask:2] 2025-09-18 05:31:29,819 RepairSession.java:270 - [repair #ba6151c0-9450-11f0-9ace-ef813b7a18da] Session completed successfully INFO [RepairJobTask:2] 2025-09-18 05:31:29,820 RepairRunnable.java:261 - Repair session ba6151c0-9450-11f0-9ace-ef813b7a18da for range [(-9083771317662873172,-9064397874770026951], (5482710683617896693,5552469124657401434]] finished INFO [RepairJobTask:2] 2025-09-18 05:31:29,823 ActiveRepairService.java:452 - [repair #ba59b0a0-9450-11f0-9ace-ef813b7a18da] Not a global repair, will not do anticompaction INFO [InternalResponseStage:8] 2025-09-18 05:31:29,827 RepairRunnable.java:343 - Repair command #1 finished in 0 seconds INFO [Repair-Task-3] 2025-09-18 05:31:37,978 RepairRunnable.java:139 - Starting repair command #2 (bf827da0-9450-11f0-9ace-ef813b7a18da), repairing keyspace reaper_db with repair options (parallelism: dc_parallel, primary range: false, incremental: false, job threads: 1, ColumnFamilies: [cluster, leader, diagnostic_event_subscription, schema_migration_leader, running_repairs, repair_run, repair_unit_v1, repair_run_by_cluster_v2, repair_schedule_by_cluster_and_keyspace, repair_run_by_cluster, running_reapers, repair_schedule_v1, percent_repaired_by_schedule, repair_run_by_unit, schema_migration, snapshot], dataCenters: [], hosts: [], # of ranges: 6, pull repair: false) INFO [Repair-Task-3] 2025-09-18 05:31:38,020 RepairSession.java:228 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] new session: will sync /10.0.0.38, /10.0.0.33, /10.0.0.22 on range [(-9209246771099019622,-9185326702906423036], (-3283111870881080472,-3234301595462477458], (3541253810191428201,3542023783901223029], (-342399068030290479,-334159928657440232], (6018591951904543460,6022995036049403353], (7023225682983239539,7029252156270852454]] for reaper_db.[repair_run, snapshot, leader, repair_unit_v1, repair_schedule_by_cluster_and_keyspace, repair_schedule_v1, schema_migration_leader, repair_run_by_cluster_v2, schema_migration, running_reapers, repair_run_by_unit, cluster, running_repairs, percent_repaired_by_schedule, repair_run_by_cluster, diagnostic_event_subscription] INFO [RepairJobTask:2] 2025-09-18 05:31:38,204 RepairJob.java:234 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for repair_run (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:2] 2025-09-18 05:31:38,204 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,217 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,217 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,226 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,226 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,247 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run from /10.0.0.38 INFO [RepairJobTask:2] 2025-09-18 05:31:38,249 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for repair_run INFO [RepairJobTask:2] 2025-09-18 05:31:38,258 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for repair_run INFO [RepairJobTask:4] 2025-09-18 05:31:38,258 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for repair_run INFO [RepairJobTask:2] 2025-09-18 05:31:38,259 RepairJob.java:143 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] repair_run is fully synced INFO [RepairJobTask:3] 2025-09-18 05:31:38,258 RepairJob.java:234 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for snapshot (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:3] 2025-09-18 05:31:38,259 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,264 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for snapshot from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,265 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,268 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for snapshot from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,268 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,291 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for snapshot from /10.0.0.38 INFO [RepairJobTask:3] 2025-09-18 05:31:38,296 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for snapshot INFO [RepairJobTask:4] 2025-09-18 05:31:38,296 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for snapshot INFO [RepairJobTask:1] 2025-09-18 05:31:38,296 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for snapshot INFO [RepairJobTask:4] 2025-09-18 05:31:38,296 RepairJob.java:143 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] snapshot is fully synced INFO [RepairJobTask:4] 2025-09-18 05:31:38,318 RepairJob.java:234 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for leader (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:4] 2025-09-18 05:31:38,318 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,322 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for leader from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,322 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,324 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for leader from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,324 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,326 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for leader from /10.0.0.38 INFO [RepairJobTask:2] 2025-09-18 05:31:38,327 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for leader INFO [RepairJobTask:4] 2025-09-18 05:31:38,327 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for leader INFO [RepairJobTask:3] 2025-09-18 05:31:38,327 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for leader INFO [RepairJobTask:4] 2025-09-18 05:31:38,327 RepairJob.java:143 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] leader is fully synced INFO [RepairJobTask:4] 2025-09-18 05:31:38,331 RepairJob.java:234 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for repair_unit_v1 (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:4] 2025-09-18 05:31:38,331 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,337 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_unit_v1 from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,337 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,342 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_unit_v1 from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,342 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,350 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_unit_v1 from /10.0.0.38 INFO [RepairJobTask:2] 2025-09-18 05:31:38,351 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for repair_unit_v1 INFO [RepairJobTask:4] 2025-09-18 05:31:38,351 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for repair_unit_v1 INFO [RepairJobTask:1] 2025-09-18 05:31:38,351 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for repair_unit_v1 INFO [RepairJobTask:4] 2025-09-18 05:31:38,352 RepairJob.java:143 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] repair_unit_v1 is fully synced INFO [RepairJobTask:4] 2025-09-18 05:31:38,373 RepairJob.java:234 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for repair_schedule_by_cluster_and_keyspace (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:4] 2025-09-18 05:31:38,373 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,379 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,379 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,382 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,382 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,386 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.38 INFO [RepairJobTask:3] 2025-09-18 05:31:38,389 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:2] 2025-09-18 05:31:38,389 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:1] 2025-09-18 05:31:38,391 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:2] 2025-09-18 05:31:38,392 RepairJob.java:143 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] repair_schedule_by_cluster_and_keyspace is fully synced INFO [RepairJobTask:2] 2025-09-18 05:31:38,401 RepairJob.java:234 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for repair_schedule_v1 (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:2] 2025-09-18 05:31:38,401 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,414 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_schedule_v1 from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,414 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,422 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_schedule_v1 from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,422 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,426 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_schedule_v1 from /10.0.0.38 INFO [RepairJobTask:4] 2025-09-18 05:31:38,427 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for repair_schedule_v1 INFO [RepairJobTask:2] 2025-09-18 05:31:38,427 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for repair_schedule_v1 INFO [RepairJobTask:5] 2025-09-18 05:31:38,431 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for repair_schedule_v1 INFO [RepairJobTask:5] 2025-09-18 05:31:38,432 RepairJob.java:143 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] repair_schedule_v1 is fully synced INFO [RepairJobTask:5] 2025-09-18 05:31:38,439 RepairJob.java:234 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for schema_migration_leader (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:5] 2025-09-18 05:31:38,440 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,447 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for schema_migration_leader from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,447 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,452 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for schema_migration_leader from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,452 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,455 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for schema_migration_leader from /10.0.0.38 INFO [RepairJobTask:1] 2025-09-18 05:31:38,456 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for schema_migration_leader INFO [RepairJobTask:1] 2025-09-18 05:31:38,456 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for schema_migration_leader INFO [RepairJobTask:1] 2025-09-18 05:31:38,456 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for schema_migration_leader INFO [RepairJobTask:1] 2025-09-18 05:31:38,456 RepairJob.java:143 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] schema_migration_leader is fully synced INFO [RepairJobTask:2] 2025-09-18 05:31:38,465 RepairJob.java:234 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for repair_run_by_cluster_v2 (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:2] 2025-09-18 05:31:38,465 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,469 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,471 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,476 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,476 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,479 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.38 INFO [RepairJobTask:2] 2025-09-18 05:31:38,482 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:2] 2025-09-18 05:31:38,482 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:2] 2025-09-18 05:31:38,482 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:2] 2025-09-18 05:31:38,482 RepairJob.java:143 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] repair_run_by_cluster_v2 is fully synced INFO [RepairJobTask:5] 2025-09-18 05:31:38,491 RepairJob.java:234 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for schema_migration (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:5] 2025-09-18 05:31:38,491 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,495 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for schema_migration from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,495 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,502 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for schema_migration from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,502 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,510 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for schema_migration from /10.0.0.38 INFO [RepairJobTask:3] 2025-09-18 05:31:38,510 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for schema_migration INFO [RepairJobTask:4] 2025-09-18 05:31:38,510 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for schema_migration INFO [RepairJobTask:2] 2025-09-18 05:31:38,511 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for schema_migration INFO [RepairJobTask:1] 2025-09-18 05:31:38,511 RepairJob.java:143 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] schema_migration is fully synced INFO [RepairJobTask:1] 2025-09-18 05:31:38,519 RepairJob.java:234 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for running_reapers (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:1] 2025-09-18 05:31:38,519 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,523 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for running_reapers from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,524 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,530 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for running_reapers from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,530 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,535 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for running_reapers from /10.0.0.38 INFO [RepairJobTask:1] 2025-09-18 05:31:38,536 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for running_reapers INFO [RepairJobTask:4] 2025-09-18 05:31:38,536 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for running_reapers INFO [RepairJobTask:3] 2025-09-18 05:31:38,536 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for running_reapers INFO [RepairJobTask:5] 2025-09-18 05:31:38,536 RepairJob.java:143 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] running_reapers is fully synced INFO [RepairJobTask:6] 2025-09-18 05:31:38,541 RepairJob.java:234 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for repair_run_by_unit (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:6] 2025-09-18 05:31:38,541 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,564 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run_by_unit from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,564 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,567 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run_by_unit from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,567 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,571 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run_by_unit from /10.0.0.38 INFO [RepairJobTask:4] 2025-09-18 05:31:38,572 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for repair_run_by_unit INFO [RepairJobTask:4] 2025-09-18 05:31:38,572 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for repair_run_by_unit INFO [RepairJobTask:4] 2025-09-18 05:31:38,572 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for repair_run_by_unit INFO [RepairJobTask:4] 2025-09-18 05:31:38,572 RepairJob.java:143 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] repair_run_by_unit is fully synced INFO [RepairJobTask:6] 2025-09-18 05:31:38,576 RepairJob.java:234 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for cluster (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:6] 2025-09-18 05:31:38,576 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,584 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for cluster from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,584 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,586 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for cluster from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,586 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,590 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for cluster from /10.0.0.38 INFO [RepairJobTask:4] 2025-09-18 05:31:38,593 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for cluster INFO [RepairJobTask:6] 2025-09-18 05:31:38,593 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for cluster INFO [RepairJobTask:3] 2025-09-18 05:31:38,593 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for cluster INFO [RepairJobTask:6] 2025-09-18 05:31:38,593 RepairJob.java:143 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] cluster is fully synced INFO [RepairJobTask:6] 2025-09-18 05:31:38,690 RepairJob.java:234 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for running_repairs (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:6] 2025-09-18 05:31:38,691 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,694 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for running_repairs from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,694 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,697 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for running_repairs from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,697 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,701 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for running_repairs from /10.0.0.38 INFO [RepairJobTask:6] 2025-09-18 05:31:38,702 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for running_repairs INFO [RepairJobTask:4] 2025-09-18 05:31:38,702 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for running_repairs INFO [RepairJobTask:5] 2025-09-18 05:31:38,704 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for running_repairs INFO [RepairJobTask:4] 2025-09-18 05:31:38,706 RepairJob.java:234 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for percent_repaired_by_schedule (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:4] 2025-09-18 05:31:38,713 RepairJob.java:257 - Validating /10.0.0.33 INFO [RepairJobTask:6] 2025-09-18 05:31:38,711 RepairJob.java:143 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] running_repairs is fully synced INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,716 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for percent_repaired_by_schedule from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,716 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,720 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for percent_repaired_by_schedule from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,720 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,723 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for percent_repaired_by_schedule from /10.0.0.38 INFO [RepairJobTask:1] 2025-09-18 05:31:38,723 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:2] 2025-09-18 05:31:38,723 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:7] 2025-09-18 05:31:38,726 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:6] 2025-09-18 05:31:38,727 RepairJob.java:234 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for repair_run_by_cluster (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:6] 2025-09-18 05:31:38,727 RepairJob.java:257 - Validating /10.0.0.33 INFO [RepairJobTask:6] 2025-09-18 05:31:38,727 RepairJob.java:143 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] percent_repaired_by_schedule is fully synced INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,729 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run_by_cluster from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,730 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,732 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run_by_cluster from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,736 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,740 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run_by_cluster from /10.0.0.38 INFO [RepairJobTask:2] 2025-09-18 05:31:38,741 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for repair_run_by_cluster INFO [RepairJobTask:2] 2025-09-18 05:31:38,741 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for repair_run_by_cluster INFO [RepairJobTask:2] 2025-09-18 05:31:38,741 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for repair_run_by_cluster INFO [RepairJobTask:2] 2025-09-18 05:31:38,741 RepairJob.java:143 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] repair_run_by_cluster is fully synced INFO [RepairJobTask:7] 2025-09-18 05:31:38,749 RepairJob.java:234 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for diagnostic_event_subscription (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:7] 2025-09-18 05:31:38,749 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,752 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for diagnostic_event_subscription from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,752 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,755 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for diagnostic_event_subscription from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,756 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:38,766 RepairSession.java:180 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Received merkle tree for diagnostic_event_subscription from /10.0.0.38 INFO [RepairJobTask:6] 2025-09-18 05:31:38,769 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for diagnostic_event_subscription INFO [RepairJobTask:5] 2025-09-18 05:31:38,769 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for diagnostic_event_subscription INFO [RepairJobTask:7] 2025-09-18 05:31:38,769 SyncTask.java:66 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for diagnostic_event_subscription INFO [RepairJobTask:5] 2025-09-18 05:31:38,769 RepairJob.java:143 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] diagnostic_event_subscription is fully synced INFO [RepairJobTask:5] 2025-09-18 05:31:38,771 RepairSession.java:270 - [repair #bf88bf30-9450-11f0-9ace-ef813b7a18da] Session completed successfully INFO [RepairJobTask:5] 2025-09-18 05:31:38,771 RepairRunnable.java:261 - Repair session bf88bf30-9450-11f0-9ace-ef813b7a18da for range [(-9209246771099019622,-9185326702906423036], (-3283111870881080472,-3234301595462477458], (3541253810191428201,3542023783901223029], (-342399068030290479,-334159928657440232], (6018591951904543460,6022995036049403353], (7023225682983239539,7029252156270852454]] finished INFO [RepairJobTask:5] 2025-09-18 05:31:38,772 ActiveRepairService.java:452 - [repair #bf827da0-9450-11f0-9ace-ef813b7a18da] Not a global repair, will not do anticompaction INFO [InternalResponseStage:8] 2025-09-18 05:31:38,778 RepairRunnable.java:343 - Repair command #2 finished in 0 seconds INFO [Repair-Task-4] 2025-09-18 05:31:39,063 RepairRunnable.java:139 - Starting repair command #3 (c027e560-9450-11f0-9ace-ef813b7a18da), repairing keyspace reaper_db with repair options (parallelism: dc_parallel, primary range: false, incremental: false, job threads: 1, ColumnFamilies: [cluster, leader, diagnostic_event_subscription, schema_migration_leader, running_repairs, repair_run, repair_unit_v1, repair_run_by_cluster_v2, repair_schedule_by_cluster_and_keyspace, repair_run_by_cluster, running_reapers, repair_schedule_v1, percent_repaired_by_schedule, repair_run_by_unit, schema_migration, snapshot], dataCenters: [], hosts: [], # of ranges: 3, pull repair: false) INFO [Repair-Task-4] 2025-09-18 05:31:39,082 RepairSession.java:228 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] new session: will sync /10.0.0.38, /10.0.0.33, /10.0.0.22 on range [(-5083698599490360897,-5062673553346518439], (1963667213076878527,1983309183220632045], (2262121229821322408,2307801754530224229]] for reaper_db.[repair_run, snapshot, leader, repair_unit_v1, repair_schedule_by_cluster_and_keyspace, repair_schedule_v1, schema_migration_leader, repair_run_by_cluster_v2, schema_migration, running_reapers, repair_run_by_unit, cluster, running_repairs, percent_repaired_by_schedule, repair_run_by_cluster, diagnostic_event_subscription] INFO [RepairJobTask:2] 2025-09-18 05:31:39,178 RepairJob.java:234 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for repair_run (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:2] 2025-09-18 05:31:39,179 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,184 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,184 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,187 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,188 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,192 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run from /10.0.0.38 INFO [RepairJobTask:2] 2025-09-18 05:31:39,195 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for repair_run INFO [RepairJobTask:4] 2025-09-18 05:31:39,195 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for repair_run INFO [RepairJobTask:2] 2025-09-18 05:31:39,196 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for repair_run INFO [RepairJobTask:2] 2025-09-18 05:31:39,196 RepairJob.java:143 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] repair_run is fully synced INFO [RepairJobTask:3] 2025-09-18 05:31:39,201 RepairJob.java:234 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for snapshot (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:3] 2025-09-18 05:31:39,201 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,205 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for snapshot from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,206 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,211 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for snapshot from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,211 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,214 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for snapshot from /10.0.0.38 INFO [RepairJobTask:2] 2025-09-18 05:31:39,214 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for snapshot INFO [RepairJobTask:2] 2025-09-18 05:31:39,223 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for snapshot INFO [RepairJobTask:2] 2025-09-18 05:31:39,224 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for snapshot INFO [RepairJobTask:2] 2025-09-18 05:31:39,224 RepairJob.java:143 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] snapshot is fully synced INFO [RepairJobTask:5] 2025-09-18 05:31:39,255 RepairJob.java:234 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for leader (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:5] 2025-09-18 05:31:39,255 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,260 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for leader from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,261 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,271 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for leader from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,271 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,282 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for leader from /10.0.0.38 INFO [RepairJobTask:2] 2025-09-18 05:31:39,285 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for leader INFO [RepairJobTask:4] 2025-09-18 05:31:39,285 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for leader INFO [RepairJobTask:1] 2025-09-18 05:31:39,285 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for leader INFO [RepairJobTask:1] 2025-09-18 05:31:39,291 RepairJob.java:234 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for repair_unit_v1 (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:1] 2025-09-18 05:31:39,292 RepairJob.java:257 - Validating /10.0.0.33 INFO [RepairJobTask:2] 2025-09-18 05:31:39,292 RepairJob.java:143 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] leader is fully synced INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,296 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_unit_v1 from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,296 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,322 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_unit_v1 from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,322 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,336 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_unit_v1 from /10.0.0.38 INFO [RepairJobTask:1] 2025-09-18 05:31:39,338 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for repair_unit_v1 INFO [RepairJobTask:2] 2025-09-18 05:31:39,339 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for repair_unit_v1 INFO [RepairJobTask:1] 2025-09-18 05:31:39,339 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for repair_unit_v1 INFO [RepairJobTask:1] 2025-09-18 05:31:39,339 RepairJob.java:143 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] repair_unit_v1 is fully synced INFO [RepairJobTask:1] 2025-09-18 05:31:39,366 RepairJob.java:234 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for repair_schedule_by_cluster_and_keyspace (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:1] 2025-09-18 05:31:39,366 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,371 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,371 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,373 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,373 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,388 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.38 INFO [RepairJobTask:1] 2025-09-18 05:31:39,391 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:2] 2025-09-18 05:31:39,390 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:3] 2025-09-18 05:31:39,398 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:2] 2025-09-18 05:31:39,399 RepairJob.java:143 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] repair_schedule_by_cluster_and_keyspace is fully synced INFO [RepairJobTask:2] 2025-09-18 05:31:39,415 RepairJob.java:234 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for repair_schedule_v1 (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:2] 2025-09-18 05:31:39,415 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,417 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_schedule_v1 from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,417 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,420 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_schedule_v1 from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,420 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,429 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_schedule_v1 from /10.0.0.38 INFO [RepairJobTask:1] 2025-09-18 05:31:39,430 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for repair_schedule_v1 INFO [RepairJobTask:3] 2025-09-18 05:31:39,430 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for repair_schedule_v1 INFO [RepairJobTask:4] 2025-09-18 05:31:39,430 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for repair_schedule_v1 INFO [RepairJobTask:3] 2025-09-18 05:31:39,430 RepairJob.java:143 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] repair_schedule_v1 is fully synced INFO [RepairJobTask:3] 2025-09-18 05:31:39,434 RepairJob.java:234 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for schema_migration_leader (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:3] 2025-09-18 05:31:39,434 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,435 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for schema_migration_leader from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,436 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,437 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for schema_migration_leader from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,442 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,445 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for schema_migration_leader from /10.0.0.38 INFO [RepairJobTask:2] 2025-09-18 05:31:39,446 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for schema_migration_leader INFO [RepairJobTask:1] 2025-09-18 05:31:39,446 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for schema_migration_leader INFO [RepairJobTask:4] 2025-09-18 05:31:39,446 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for schema_migration_leader INFO [RepairJobTask:1] 2025-09-18 05:31:39,446 RepairJob.java:143 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] schema_migration_leader is fully synced INFO [RepairJobTask:1] 2025-09-18 05:31:39,452 RepairJob.java:234 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for repair_run_by_cluster_v2 (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:1] 2025-09-18 05:31:39,453 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,459 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,460 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,470 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,470 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,472 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.38 INFO [RepairJobTask:3] 2025-09-18 05:31:39,472 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:1] 2025-09-18 05:31:39,473 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:4] 2025-09-18 05:31:39,473 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:1] 2025-09-18 05:31:39,473 RepairJob.java:143 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] repair_run_by_cluster_v2 is fully synced INFO [RepairJobTask:1] 2025-09-18 05:31:39,478 RepairJob.java:234 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for schema_migration (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:1] 2025-09-18 05:31:39,478 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,480 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for schema_migration from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,480 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,482 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for schema_migration from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,482 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,484 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for schema_migration from /10.0.0.38 INFO [RepairJobTask:5] 2025-09-18 05:31:39,487 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for schema_migration INFO [RepairJobTask:1] 2025-09-18 05:31:39,489 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for schema_migration INFO [RepairJobTask:6] 2025-09-18 05:31:39,490 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for schema_migration INFO [RepairJobTask:1] 2025-09-18 05:31:39,490 RepairJob.java:143 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] schema_migration is fully synced INFO [RepairJobTask:6] 2025-09-18 05:31:39,496 RepairJob.java:234 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for running_reapers (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:6] 2025-09-18 05:31:39,496 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,503 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for running_reapers from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,503 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,509 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for running_reapers from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,509 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,514 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for running_reapers from /10.0.0.38 INFO [RepairJobTask:1] 2025-09-18 05:31:39,515 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for running_reapers INFO [RepairJobTask:5] 2025-09-18 05:31:39,515 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for running_reapers INFO [RepairJobTask:4] 2025-09-18 05:31:39,515 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for running_reapers INFO [RepairJobTask:5] 2025-09-18 05:31:39,515 RepairJob.java:143 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] running_reapers is fully synced INFO [RepairJobTask:6] 2025-09-18 05:31:39,521 RepairJob.java:234 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for repair_run_by_unit (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:6] 2025-09-18 05:31:39,522 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,529 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run_by_unit from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,529 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,531 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run_by_unit from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,531 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,537 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run_by_unit from /10.0.0.38 INFO [RepairJobTask:4] 2025-09-18 05:31:39,538 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for repair_run_by_unit INFO [RepairJobTask:3] 2025-09-18 05:31:39,538 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for repair_run_by_unit INFO [RepairJobTask:5] 2025-09-18 05:31:39,538 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for repair_run_by_unit INFO [RepairJobTask:3] 2025-09-18 05:31:39,538 RepairJob.java:143 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] repair_run_by_unit is fully synced INFO [RepairJobTask:3] 2025-09-18 05:31:39,543 RepairJob.java:234 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for cluster (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:3] 2025-09-18 05:31:39,546 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,554 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for cluster from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,554 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,556 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for cluster from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,558 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,560 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for cluster from /10.0.0.38 INFO [RepairJobTask:1] 2025-09-18 05:31:39,561 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for cluster INFO [RepairJobTask:5] 2025-09-18 05:31:39,561 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for cluster INFO [RepairJobTask:2] 2025-09-18 05:31:39,560 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for cluster INFO [RepairJobTask:6] 2025-09-18 05:31:39,561 RepairJob.java:143 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] cluster is fully synced INFO [RepairJobTask:6] 2025-09-18 05:31:39,615 RepairJob.java:234 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for running_repairs (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:6] 2025-09-18 05:31:39,616 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,619 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for running_repairs from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,619 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,623 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for running_repairs from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,623 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,624 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for running_repairs from /10.0.0.38 INFO [RepairJobTask:5] 2025-09-18 05:31:39,624 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for running_repairs INFO [RepairJobTask:6] 2025-09-18 05:31:39,624 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for running_repairs INFO [RepairJobTask:4] 2025-09-18 05:31:39,625 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for running_repairs INFO [RepairJobTask:1] 2025-09-18 05:31:39,625 RepairJob.java:143 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] running_repairs is fully synced INFO [RepairJobTask:7] 2025-09-18 05:31:39,635 RepairJob.java:234 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for percent_repaired_by_schedule (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:7] 2025-09-18 05:31:39,636 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,639 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for percent_repaired_by_schedule from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,639 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,641 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for percent_repaired_by_schedule from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,641 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,644 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for percent_repaired_by_schedule from /10.0.0.38 INFO [RepairJobTask:1] 2025-09-18 05:31:39,644 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:1] 2025-09-18 05:31:39,644 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:3] 2025-09-18 05:31:39,644 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:4] 2025-09-18 05:31:39,645 RepairJob.java:143 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] percent_repaired_by_schedule is fully synced INFO [RepairJobTask:4] 2025-09-18 05:31:39,649 RepairJob.java:234 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for repair_run_by_cluster (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:4] 2025-09-18 05:31:39,650 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,654 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run_by_cluster from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,654 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,656 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run_by_cluster from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,656 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,661 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for repair_run_by_cluster from /10.0.0.38 INFO [RepairJobTask:1] 2025-09-18 05:31:39,661 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for repair_run_by_cluster INFO [RepairJobTask:5] 2025-09-18 05:31:39,661 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for repair_run_by_cluster INFO [RepairJobTask:2] 2025-09-18 05:31:39,661 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for repair_run_by_cluster INFO [RepairJobTask:6] 2025-09-18 05:31:39,662 RepairJob.java:143 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] repair_run_by_cluster is fully synced INFO [RepairJobTask:6] 2025-09-18 05:31:39,667 RepairJob.java:234 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Requesting merkle trees for diagnostic_event_subscription (to [/10.0.0.33, /10.0.0.22, /10.0.0.38]) INFO [RepairJobTask:6] 2025-09-18 05:31:39,667 RepairJob.java:257 - Validating /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,669 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for diagnostic_event_subscription from /10.0.0.33 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,669 RepairJob.java:270 - Validating /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,672 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for diagnostic_event_subscription from /10.0.0.22 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,672 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-09-18 05:31:39,674 RepairSession.java:180 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Received merkle tree for diagnostic_event_subscription from /10.0.0.38 INFO [RepairJobTask:3] 2025-09-18 05:31:39,674 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.38 are consistent for diagnostic_event_subscription INFO [RepairJobTask:5] 2025-09-18 05:31:39,674 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.22 and /10.0.0.38 are consistent for diagnostic_event_subscription INFO [RepairJobTask:6] 2025-09-18 05:31:39,675 SyncTask.java:66 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Endpoints /10.0.0.33 and /10.0.0.22 are consistent for diagnostic_event_subscription INFO [RepairJobTask:5] 2025-09-18 05:31:39,677 RepairJob.java:143 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] diagnostic_event_subscription is fully synced INFO [RepairJobTask:5] 2025-09-18 05:31:39,677 RepairSession.java:270 - [repair #c02af2a0-9450-11f0-9ace-ef813b7a18da] Session completed successfully INFO [RepairJobTask:5] 2025-09-18 05:31:39,678 RepairRunnable.java:261 - Repair session c02af2a0-9450-11f0-9ace-ef813b7a18da for range [(-5083698599490360897,-5062673553346518439], (1963667213076878527,1983309183220632045], (2262121229821322408,2307801754530224229]] finished INFO [RepairJobTask:5] 2025-09-18 05:31:39,678 ActiveRepairService.java:452 - [repair #c027e560-9450-11f0-9ace-ef813b7a18da] Not a global repair, will not do anticompaction INFO [InternalResponseStage:8] 2025-09-18 05:31:39,681 RepairRunnable.java:343 - Repair command #3 finished in 0 seconds