INFO [main] 2025-07-24 06:33:30,886 YamlConfigurationLoader.java:89 - Configuration location: file:/etc/cassandra/cassandra.yaml INFO [main] 2025-07-24 06:33:31,270 Config.java:495 - Node configuration:[allocate_tokens_for_keyspace=null; authenticator=AllowAllAuthenticator; authorizer=AllowAllAuthorizer; auto_bootstrap=true; auto_snapshot=true; back_pressure_enabled=false; back_pressure_strategy=org.apache.cassandra.net.RateBasedBackPressure{high_ratio=0.9, factor=5, flow=FAST}; batch_size_fail_threshold_in_kb=50; batch_size_warn_threshold_in_kb=5; batchlog_replay_throttle_in_kb=1024; broadcast_address=10.0.0.254; broadcast_rpc_address=10.0.0.254; buffer_pool_use_heap_if_exhausted=true; cas_contention_timeout_in_ms=1000; cdc_enabled=false; cdc_free_space_check_interval_ms=250; cdc_raw_directory=null; cdc_total_space_in_mb=0; client_encryption_options=; cluster_name=contrail_database; column_index_cache_size_in_kb=2; column_index_size_in_kb=64; commit_failure_policy=stop; commitlog_compression=null; commitlog_directory=/var/lib/cassandra/commitlog; commitlog_max_compression_buffers_in_pool=3; commitlog_periodic_queue_size=-1; commitlog_segment_size_in_mb=32; commitlog_sync=periodic; commitlog_sync_batch_window_in_ms=NaN; commitlog_sync_period_in_ms=10000; commitlog_total_space_in_mb=null; compaction_large_partition_warning_threshold_mb=100; compaction_throughput_mb_per_sec=256; concurrent_compactors=4; concurrent_counter_writes=32; concurrent_materialized_view_writes=32; concurrent_reads=64; concurrent_replicates=null; concurrent_writes=64; counter_cache_keys_to_save=2147483647; counter_cache_save_period=7200; counter_cache_size_in_mb=null; counter_write_request_timeout_in_ms=5000; credentials_cache_max_entries=1000; credentials_update_interval_in_ms=-1; credentials_validity_in_ms=2000; cross_node_timeout=false; data_file_directories=[Ljava.lang.String;@6b19b79; disk_access_mode=auto; disk_failure_policy=stop; disk_optimization_estimate_percentile=0.95; disk_optimization_page_cross_chance=0.1; disk_optimization_strategy=ssd; dynamic_snitch=true; dynamic_snitch_badness_threshold=0.1; dynamic_snitch_reset_interval_in_ms=600000; dynamic_snitch_update_interval_in_ms=100; enable_materialized_views=true; enable_scripted_user_defined_functions=false; enable_user_defined_functions=false; enable_user_defined_functions_threads=true; encryption_options=null; endpoint_snitch=SimpleSnitch; file_cache_round_up=null; file_cache_size_in_mb=null; gc_log_threshold_in_ms=200; gc_warn_threshold_in_ms=1000; hinted_handoff_disabled_datacenters=[]; hinted_handoff_enabled=true; hinted_handoff_throttle_in_kb=1024; hints_compression=null; hints_directory=null; hints_flush_period_in_ms=10000; incremental_backups=false; index_interval=null; index_summary_capacity_in_mb=null; index_summary_resize_interval_in_minutes=60; initial_token=null; inter_dc_stream_throughput_outbound_megabits_per_sec=200; inter_dc_tcp_nodelay=false; internode_authenticator=null; internode_compression=dc; internode_recv_buff_size_in_bytes=0; internode_send_buff_size_in_bytes=0; key_cache_keys_to_save=2147483647; key_cache_save_period=14400; key_cache_size_in_mb=null; listen_address=10.0.0.254; listen_interface=null; listen_interface_prefer_ipv6=false; listen_on_broadcast_address=false; max_hint_window_in_ms=10800000; max_hints_delivery_threads=2; max_hints_file_size_in_mb=128; max_mutation_size_in_kb=null; max_streaming_retries=3; max_value_size_in_mb=256; memtable_allocation_type=offheap_objects; memtable_cleanup_threshold=null; memtable_flush_writers=4; memtable_heap_space_in_mb=null; memtable_offheap_space_in_mb=null; min_free_space_per_drive_in_mb=50; native_transport_max_concurrent_connections=-1; native_transport_max_concurrent_connections_per_ip=-1; native_transport_max_frame_size_in_mb=256; native_transport_max_threads=128; native_transport_port=9042; native_transport_port_ssl=null; num_tokens=256; otc_backlog_expiration_interval_ms=200; otc_coalescing_enough_coalesced_messages=8; otc_coalescing_strategy=DISABLED; otc_coalescing_window_us=200; partitioner=org.apache.cassandra.dht.Murmur3Partitioner; permissions_cache_max_entries=1000; permissions_update_interval_in_ms=-1; permissions_validity_in_ms=2000; phi_convict_threshold=8.0; prepared_statements_cache_size_mb=null; range_request_timeout_in_ms=10000; read_request_timeout_in_ms=5000; request_scheduler=org.apache.cassandra.scheduler.NoScheduler; request_scheduler_id=null; request_scheduler_options=null; request_timeout_in_ms=10000; role_manager=CassandraRoleManager; roles_cache_max_entries=1000; roles_update_interval_in_ms=-1; roles_validity_in_ms=2000; row_cache_class_name=org.apache.cassandra.cache.OHCProvider; row_cache_keys_to_save=2147483647; row_cache_save_period=0; row_cache_size_in_mb=0; rpc_address=10.0.0.254; rpc_interface=null; rpc_interface_prefer_ipv6=false; rpc_keepalive=true; rpc_listen_backlog=50; rpc_max_threads=2147483647; rpc_min_threads=16; rpc_port=9160; rpc_recv_buff_size_in_bytes=null; rpc_send_buff_size_in_bytes=null; rpc_server_type=sync; saved_caches_directory=/var/lib/cassandra/saved_caches; seed_provider=org.apache.cassandra.locator.SimpleSeedProvider{seeds=10.0.0.254,10.0.0.38}; server_encryption_options=; slow_query_log_timeout_in_ms=500; snapshot_before_compaction=false; ssl_storage_port=7001; sstable_preemptive_open_interval_in_mb=50; start_native_transport=true; start_rpc=true; storage_port=7000; stream_throughput_outbound_megabits_per_sec=200; streaming_keep_alive_period_in_secs=300; streaming_socket_timeout_in_ms=86400000; thrift_framed_transport_size_in_mb=15; thrift_max_message_length_in_mb=16; thrift_prepared_statements_cache_size_mb=null; tombstone_failure_threshold=100000; tombstone_warn_threshold=1000; tracetype_query_ttl=86400; tracetype_repair_ttl=604800; transparent_data_encryption_options=org.apache.cassandra.config.TransparentDataEncryptionOptions@2a32de6c; trickle_fsync=false; trickle_fsync_interval_in_kb=10240; truncate_request_timeout_in_ms=60000; unlogged_batch_across_partitions_warn_threshold=10; user_defined_function_fail_timeout=1500; user_defined_function_warn_timeout=500; user_function_timeout_policy=die; windows_timer_interval=1; write_request_timeout_in_ms=2000] INFO [main] 2025-07-24 06:33:31,271 DatabaseDescriptor.java:367 - DiskAccessMode 'auto' determined to be mmap, indexAccessMode is mmap INFO [main] 2025-07-24 06:33:31,271 DatabaseDescriptor.java:425 - Global memtable on-heap threshold is enabled at 502MB INFO [main] 2025-07-24 06:33:31,271 DatabaseDescriptor.java:429 - Global memtable off-heap threshold is enabled at 502MB INFO [main] 2025-07-24 06:33:31,321 RateBasedBackPressure.java:123 - Initialized back-pressure with high ratio: 0.9, factor: 5, flow: FAST, window size: 2000. INFO [main] 2025-07-24 06:33:31,322 DatabaseDescriptor.java:729 - Back-pressure is disabled with strategy org.apache.cassandra.net.RateBasedBackPressure{high_ratio=0.9, factor=5, flow=FAST}. INFO [main] 2025-07-24 06:33:31,615 JMXServerUtils.java:246 - Configured JMX server at: service:jmx:rmi://0.0.0.0/jndi/rmi://0.0.0.0:7201/jmxrmi INFO [main] 2025-07-24 06:33:31,629 CassandraDaemon.java:473 - Hostname: cn-jenkins-deploy-platform-ansible-os-3714-1. INFO [main] 2025-07-24 06:33:31,630 CassandraDaemon.java:480 - JVM vendor/version: OpenJDK 64-Bit Server VM/1.8.0_322 INFO [main] 2025-07-24 06:33:31,631 CassandraDaemon.java:481 - Heap size: 984.000MiB/1.961GiB INFO [main] 2025-07-24 06:33:31,631 CassandraDaemon.java:486 - Code Cache Non-heap memory: init = 2555904(2496K) used = 4227008(4127K) committed = 4259840(4160K) max = 251658240(245760K) INFO [main] 2025-07-24 06:33:31,632 CassandraDaemon.java:486 - Metaspace Non-heap memory: init = 0(0K) used = 19672448(19211K) committed = 20316160(19840K) max = -1(-1K) INFO [main] 2025-07-24 06:33:31,633 CassandraDaemon.java:486 - Compressed Class Space Non-heap memory: init = 0(0K) used = 2304432(2250K) committed = 2490368(2432K) max = 1073741824(1048576K) INFO [main] 2025-07-24 06:33:31,634 CassandraDaemon.java:486 - Par Eden Space Heap memory: init = 335544320(327680K) used = 93995472(91792K) committed = 335544320(327680K) max = 335544320(327680K) INFO [main] 2025-07-24 06:33:31,643 CassandraDaemon.java:486 - Par Survivor Space Heap memory: init = 41943040(40960K) used = 0(0K) committed = 41943040(40960K) max = 41943040(40960K) INFO [main] 2025-07-24 06:33:31,644 CassandraDaemon.java:486 - CMS Old Gen Heap memory: init = 654311424(638976K) used = 0(0K) committed = 654311424(638976K) max = 1728053248(1687552K) INFO [main] 2025-07-24 06:33:31,645 CassandraDaemon.java:488 - Classpath: /opt/cassandra/conf:/opt/cassandra/build/classes/main:/opt/cassandra/build/classes/thrift:/opt/cassandra/lib/airline-0.6.jar:/opt/cassandra/lib/antlr-runtime-3.5.2.jar:/opt/cassandra/lib/apache-cassandra-3.11.3.jar:/opt/cassandra/lib/apache-cassandra-thrift-3.11.3.jar:/opt/cassandra/lib/asm-5.0.4.jar:/opt/cassandra/lib/caffeine-2.2.6.jar:/opt/cassandra/lib/cassandra-driver-core-3.0.1-shaded.jar:/opt/cassandra/lib/commons-cli-1.1.jar:/opt/cassandra/lib/commons-codec-1.9.jar:/opt/cassandra/lib/commons-lang3-3.1.jar:/opt/cassandra/lib/commons-math3-3.2.jar:/opt/cassandra/lib/compress-lzf-0.8.4.jar:/opt/cassandra/lib/concurrentlinkedhashmap-lru-1.4.jar:/opt/cassandra/lib/concurrent-trees-2.4.0.jar:/opt/cassandra/lib/disruptor-3.0.1.jar:/opt/cassandra/lib/ecj-4.4.2.jar:/opt/cassandra/lib/guava-18.0.jar:/opt/cassandra/lib/HdrHistogram-2.1.9.jar:/opt/cassandra/lib/high-scale-lib-1.0.6.jar:/opt/cassandra/lib/hppc-0.5.4.jar:/opt/cassandra/lib/jackson-core-asl-1.9.13.jar:/opt/cassandra/lib/jackson-mapper-asl-1.9.13.jar:/opt/cassandra/lib/jamm-0.3.0.jar:/opt/cassandra/lib/javax.inject.jar:/opt/cassandra/lib/jbcrypt-0.3m.jar:/opt/cassandra/lib/jcl-over-slf4j-1.7.7.jar:/opt/cassandra/lib/jctools-core-1.2.1.jar:/opt/cassandra/lib/jflex-1.6.0.jar:/opt/cassandra/lib/jna-4.2.2.jar:/opt/cassandra/lib/joda-time-2.4.jar:/opt/cassandra/lib/json-simple-1.1.jar:/opt/cassandra/lib/jstackjunit-0.0.1.jar:/opt/cassandra/lib/libthrift-0.13.0.jar:/opt/cassandra/lib/log4j-over-slf4j-1.7.7.jar:/opt/cassandra/lib/logback-classic-1.2.9.jar:/opt/cassandra/lib/logback-core-1.2.9.jar:/opt/cassandra/lib/lz4-1.3.0.jar:/opt/cassandra/lib/metrics-core-3.1.5.jar:/opt/cassandra/lib/metrics-jvm-3.1.5.jar:/opt/cassandra/lib/metrics-logback-3.1.5.jar:/opt/cassandra/lib/netty-all-4.1.39.Final.jar:/opt/cassandra/lib/ohc-core-0.4.4.jar:/opt/cassandra/lib/ohc-core-j8-0.4.4.jar:/opt/cassandra/lib/reporter-config3-3.0.3.jar:/opt/cassandra/lib/reporter-config-base-3.0.3.jar:/opt/cassandra/lib/sigar-1.6.4.jar:/opt/cassandra/lib/slf4j-api-1.7.7.jar:/opt/cassandra/lib/snakeyaml-1.11.jar:/opt/cassandra/lib/snappy-java-1.1.1.7.jar:/opt/cassandra/lib/snowball-stemmer-1.3.0.581.1.jar:/opt/cassandra/lib/ST4-4.0.8.jar:/opt/cassandra/lib/stream-2.5.2.jar:/opt/cassandra/lib/thrift-server-0.3.7.jar:/opt/cassandra/lib/jsr223/*/*.jar:/opt/cassandra/lib/jamm-0.3.0.jar INFO [main] 2025-07-24 06:33:31,647 CassandraDaemon.java:490 - JVM Arguments: [-Xloggc:/opt/cassandra/logs/gc.log, -ea, -XX:+UseThreadPriorities, -XX:ThreadPriorityPolicy=42, -XX:+HeapDumpOnOutOfMemoryError, -Xss256k, -XX:StringTableSize=1000003, -XX:+AlwaysPreTouch, -XX:-UseBiasedLocking, -XX:+UseTLAB, -XX:+ResizeTLAB, -XX:+UseNUMA, -XX:+PerfDisableSharedMem, -Djava.net.preferIPv4Stack=true, -Xms1g, -Xmx2g, -XX:+UseParNewGC, -XX:+UseConcMarkSweepGC, -XX:+CMSParallelRemarkEnabled, -XX:SurvivorRatio=8, -XX:MaxTenuringThreshold=1, -XX:CMSInitiatingOccupancyFraction=75, -XX:+UseCMSInitiatingOccupancyOnly, -XX:CMSWaitDuration=10000, -XX:+CMSParallelInitialMarkEnabled, -XX:+CMSEdenChunksRecordAlways, -XX:+CMSClassUnloadingEnabled, -XX:+PrintGCDetails, -XX:+PrintGCDateStamps, -XX:+PrintHeapAtGC, -XX:+PrintTenuringDistribution, -XX:+PrintGCApplicationStoppedTime, -XX:+PrintPromotionFailure, -XX:+UseGCLogFileRotation, -XX:NumberOfGCLogFiles=10, -XX:GCLogFileSize=10M, -Xmn400M, -XX:+UseCondCardMark, -XX:CompileCommandFile=/opt/cassandra/conf/hotspot_compiler, -javaagent:/opt/cassandra/lib/jamm-0.3.0.jar, -Dcassandra.jmx.remote.port=7199, -Dcom.sun.management.jmxremote.rmi.port=7199, -Dcom.sun.management.jmxremote.authenticate=true, -Dcom.sun.management.jmxremote.password.file=/etc/cassandra/jmxremote.password, -Djava.library.path=/opt/cassandra/lib/sigar-bin, -Dcassandra.rpc_port=9161, -Dcassandra.native_transport_port=9041, -Dcassandra.ssl_storage_port=7013, -Dcassandra.storage_port=7012, -Dcassandra.jmx.local.port=7201, -Dcom.sun.management.jmxremote.access.file=/etc/cassandra/jmxremote.access, -Dcassandra.jmx.remote.port=7201, -Dcom.sun.management.jmxremote.rmi.port=7201, -Dcassandra.libjemalloc=/usr/lib64/libjemalloc.so.1, -XX:OnOutOfMemoryError=kill -9 %p, -Dlogback.configurationFile=logback.xml, -Dcassandra.logdir=/opt/cassandra/logs, -Dcassandra.storagedir=/opt/cassandra/data, -Dcassandra-foreground=yes] WARN [main] 2025-07-24 06:33:31,726 NativeLibrary.java:187 - Unable to lock JVM memory (ENOMEM). This can result in part of the JVM being swapped out, especially with mmapped I/O enabled. Increase RLIMIT_MEMLOCK or run Cassandra as root. INFO [main] 2025-07-24 06:33:31,727 StartupChecks.java:140 - jemalloc seems to be preloaded from /usr/lib64/libjemalloc.so.1 INFO [main] 2025-07-24 06:33:31,727 StartupChecks.java:176 - JMX is enabled to receive remote connections on port: 7201 INFO [main] 2025-07-24 06:33:31,728 SigarLibrary.java:44 - Initializing SIGAR library INFO [main] 2025-07-24 06:33:31,739 SigarLibrary.java:180 - Checked OS settings and found them configured for optimal performance. WARN [main] 2025-07-24 06:33:31,742 StartupChecks.java:311 - Maximum number of memory map areas per process (vm.max_map_count) 128960 is too low, recommended value: 1048575, you can change it with sysctl. WARN [main] 2025-07-24 06:33:31,752 StartupChecks.java:332 - Directory /var/lib/cassandra/commitlog doesn't exist WARN [main] 2025-07-24 06:33:31,753 StartupChecks.java:332 - Directory /var/lib/cassandra/saved_caches doesn't exist WARN [main] 2025-07-24 06:33:31,758 StartupChecks.java:332 - Directory /opt/cassandra/data/hints doesn't exist INFO [main] 2025-07-24 06:33:31,833 QueryProcessor.java:116 - Initialized prepared statement caches with 10 MB (native) and 10 MB (Thrift) INFO [main] 2025-07-24 06:33:32,615 ColumnFamilyStore.java:411 - Initializing system.IndexInfo INFO [main] 2025-07-24 06:33:34,986 ColumnFamilyStore.java:411 - Initializing system.batches INFO [main] 2025-07-24 06:33:34,996 ColumnFamilyStore.java:411 - Initializing system.paxos INFO [main] 2025-07-24 06:33:35,013 ColumnFamilyStore.java:411 - Initializing system.local INFO [main] 2025-07-24 06:33:35,026 ColumnFamilyStore.java:411 - Initializing system.peers INFO [main] 2025-07-24 06:33:35,031 ColumnFamilyStore.java:411 - Initializing system.peer_events INFO [main] 2025-07-24 06:33:35,036 ColumnFamilyStore.java:411 - Initializing system.range_xfers INFO [main] 2025-07-24 06:33:35,040 ColumnFamilyStore.java:411 - Initializing system.compaction_history INFO [main] 2025-07-24 06:33:35,044 ColumnFamilyStore.java:411 - Initializing system.sstable_activity INFO [main] 2025-07-24 06:33:35,051 ColumnFamilyStore.java:411 - Initializing system.size_estimates INFO [main] 2025-07-24 06:33:35,056 ColumnFamilyStore.java:411 - Initializing system.available_ranges INFO [main] 2025-07-24 06:33:35,066 ColumnFamilyStore.java:411 - Initializing system.transferred_ranges INFO [main] 2025-07-24 06:33:35,083 ColumnFamilyStore.java:411 - Initializing system.views_builds_in_progress INFO [main] 2025-07-24 06:33:35,096 ColumnFamilyStore.java:411 - Initializing system.built_views INFO [main] 2025-07-24 06:33:35,107 ColumnFamilyStore.java:411 - Initializing system.hints INFO [main] 2025-07-24 06:33:35,115 ColumnFamilyStore.java:411 - Initializing system.batchlog INFO [main] 2025-07-24 06:33:35,119 ColumnFamilyStore.java:411 - Initializing system.prepared_statements INFO [main] 2025-07-24 06:33:35,124 ColumnFamilyStore.java:411 - Initializing system.schema_keyspaces INFO [main] 2025-07-24 06:33:35,131 ColumnFamilyStore.java:411 - Initializing system.schema_columnfamilies INFO [main] 2025-07-24 06:33:35,135 ColumnFamilyStore.java:411 - Initializing system.schema_columns INFO [main] 2025-07-24 06:33:35,144 ColumnFamilyStore.java:411 - Initializing system.schema_triggers INFO [main] 2025-07-24 06:33:35,148 ColumnFamilyStore.java:411 - Initializing system.schema_usertypes INFO [main] 2025-07-24 06:33:35,154 ColumnFamilyStore.java:411 - Initializing system.schema_functions INFO [main] 2025-07-24 06:33:35,160 ColumnFamilyStore.java:411 - Initializing system.schema_aggregates INFO [main] 2025-07-24 06:33:35,162 ViewManager.java:137 - Not submitting build tasks for views in keyspace system as storage service is not initialized INFO [main] 2025-07-24 06:33:35,286 ApproximateTime.java:44 - Scheduling approximate time-check task with a precision of 10 milliseconds INFO [main] 2025-07-24 06:33:35,358 ColumnFamilyStore.java:411 - Initializing system_schema.keyspaces INFO [main] 2025-07-24 06:33:35,366 ColumnFamilyStore.java:411 - Initializing system_schema.tables INFO [main] 2025-07-24 06:33:35,375 ColumnFamilyStore.java:411 - Initializing system_schema.columns INFO [main] 2025-07-24 06:33:35,383 ColumnFamilyStore.java:411 - Initializing system_schema.triggers INFO [main] 2025-07-24 06:33:35,391 ColumnFamilyStore.java:411 - Initializing system_schema.dropped_columns INFO [main] 2025-07-24 06:33:35,401 ColumnFamilyStore.java:411 - Initializing system_schema.views INFO [main] 2025-07-24 06:33:35,408 ColumnFamilyStore.java:411 - Initializing system_schema.types INFO [main] 2025-07-24 06:33:35,417 ColumnFamilyStore.java:411 - Initializing system_schema.functions INFO [main] 2025-07-24 06:33:35,430 ColumnFamilyStore.java:411 - Initializing system_schema.aggregates INFO [main] 2025-07-24 06:33:35,444 ColumnFamilyStore.java:411 - Initializing system_schema.indexes INFO [main] 2025-07-24 06:33:35,453 ViewManager.java:137 - Not submitting build tasks for views in keyspace system_schema as storage service is not initialized INFO [MemtableFlushWriter:1] 2025-07-24 06:33:36,547 CacheService.java:112 - Initializing key cache with capacity of 49 MBs. INFO [MemtableFlushWriter:1] 2025-07-24 06:33:36,553 CacheService.java:134 - Initializing row cache with capacity of 0 MBs INFO [MemtableFlushWriter:1] 2025-07-24 06:33:36,669 CacheService.java:163 - Initializing counter cache with capacity of 24 MBs INFO [MemtableFlushWriter:1] 2025-07-24 06:33:36,672 CacheService.java:174 - Scheduling counter cache save to every 7200 seconds (going to save all keys). INFO [CompactionExecutor:1] 2025-07-24 06:33:36,999 BufferPool.java:230 - Global buffer pool is enabled, when pool is exhausted (max is 502.000MiB) it will allocate on heap INFO [main] 2025-07-24 06:33:37,161 StorageService.java:600 - Populating token metadata from system tables INFO [main] 2025-07-24 06:33:37,252 StorageService.java:607 - Token metadata: INFO [pool-4-thread-1] 2025-07-24 06:33:37,345 AutoSavingCache.java:174 - Completed loading (12 ms; 1 keys) KeyCache cache INFO [main] 2025-07-24 06:33:37,380 CommitLog.java:152 - No commitlog files found; skipping replay INFO [main] 2025-07-24 06:33:37,381 StorageService.java:600 - Populating token metadata from system tables INFO [main] 2025-07-24 06:33:37,433 StorageService.java:607 - Token metadata: INFO [main] 2025-07-24 06:33:37,624 QueryProcessor.java:163 - Preloaded 0 prepared statements INFO [main] 2025-07-24 06:33:37,625 StorageService.java:618 - Cassandra version: 3.11.3 INFO [main] 2025-07-24 06:33:37,625 StorageService.java:619 - Thrift API version: 20.1.0 INFO [main] 2025-07-24 06:33:37,625 StorageService.java:620 - CQL supported versions: 3.4.4 (default: 3.4.4) INFO [main] 2025-07-24 06:33:37,626 StorageService.java:622 - Native protocol supported versions: 3/v3, 4/v4, 5/v5-beta (default: 4/v4) INFO [main] 2025-07-24 06:33:37,667 IndexSummaryManager.java:85 - Initializing index summary manager with a memory pool size of 49 MB and a resize interval of 60 minutes INFO [main] 2025-07-24 06:33:37,688 MessagingService.java:761 - Starting Messaging Service on /10.0.0.254:7012 (ens3) WARN [main] 2025-07-24 06:33:37,695 SystemKeyspace.java:1087 - No host ID found, created d21eeb5d-15a3-42c3-8f2f-d4fc4392fe21 (Note: This should happen exactly once per node). INFO [main] 2025-07-24 06:33:37,763 OutboundTcpConnection.java:108 - OutboundTcpConnection using coalescing strategy DISABLED INFO [HANDSHAKE-/10.0.0.38] 2025-07-24 06:33:37,830 OutboundTcpConnection.java:561 - Handshaking version with /10.0.0.38 INFO [HANDSHAKE-/10.0.0.50] 2025-07-24 06:33:37,970 OutboundTcpConnection.java:561 - Handshaking version with /10.0.0.50 INFO [main] 2025-07-24 06:33:38,819 StorageService.java:550 - Unable to gossip with any peers but continuing anyway since node is in its own seed list INFO [main] 2025-07-24 06:33:38,886 StorageService.java:704 - Loading persisted ring state INFO [main] 2025-07-24 06:33:38,903 StorageService.java:822 - Starting up server gossip INFO [main] 2025-07-24 06:33:39,144 StorageService.java:883 - This node will not auto bootstrap because it is configured to be a seed node. INFO [main] 2025-07-24 06:33:39,170 BootStrapper.java:228 - Generated random tokens. tokens are [-2380903637619073677, 5798173095790140686, -9214974424005107887, -356101694007609684, -2169814430967593721, 3822976179151997106, -6300969214362159072, -1526568424569693966, -2281910641657107235, 2253659061149937436, 5674588708077039055, 5210545058963264753, -6078160011078791248, -4906946210851706464, -6875955105020631945, 2801907427055709738, -1566512235200285266, -6001506251782148027, -4042217983364620029, -878106028116063246, 6422907559987374954, 8528752761094383389, -6120269011740111979, 7424828510837681589, -7918471475366111071, 3214263329102498935, 2124076021426425264, 5883872083712430195, -2819761816560715311, 9130044659852930137, 8323820669870190290, -5528399886071335802, 6758320479121489691, 6349411192502667774, 5425932069424685688, 953856701114121747, 883692109801098422, 4484351423374867871, 6876739400761999366, -3543368764333444668, 1810283819127030121, 8739318587616753452, -8181183169414006319, 3631382270898108493, -6220763695671486370, -5895629196514246564, 4767251379733088862, -781425316141887880, -5477423444602818617, -2674794296584888671, 6049020662136055301, 7810931003621179279, 610402726800367351, -794126515646826847, -7318872385044810081, 6619760434128124651, 4762245539925730583, -8704259484084181733, -4060753261069659453, -7650994337425184860, -2274524779930525524, -8934709197242634281, 7883754972648500365, 5552167499886392255, -2899913137496079876, 6907782391951312165, 4872670630382777473, 1836249025498465846, 1578655764913464537, -7012595663486618148, 5724165183228208402, -7471356289563632374, 6616574324981538456, -8546752834222015148, 861556380417746245, 6904821472753266991, -5154881281679339510, 891076370727160075, 3715881836314624284, 431536770245828889, -4732470862920118843, -721521954005108465, 3365192772816172713, -447340833209014350, -3878534752191500006, -564073279801513788, 372336364978940869, 3120628635273439109, 6026198266196935397, 2899395773676179130, -6826611487310960167, -2696291483629537723, -5562840488161224037, 6055433353200936123, -4238581110922972508, -8991036155887496588, 7790805792559227305, -5439316967732927070, 4277924291787267445, 1444952535838513127, -9142752316372498013, -6587944049123340745, 2191785578294424479, 1862596318193765887, 7321606276054746299, 193758348444948332, -8517664143011380902, -2450261117370903193, -9138202464693944584, -6797087424061630492, -6708834105207609133, 7138170488140418614, 7056702475758715799, -615413880655739322, 8606929898536881278, 3853452655386559095, 8623676962966911141, -3176877137333374973, 6427969223744857189, -1928538515432425370, 3885583054259462764, -2940287485871702302, -7911021323759761363, -6770983012924104590, -2532289513094084172, -8463175969935808900, -2164387514720836910, -2772785064730899467, 8240755233308558787, 2406735802826058580, -8698920969505912922, 2642028224764384596, -8407708505571574165, -6530181907875014184, -5901878627400801267, -8698228066245581810, -4397648956175359535, 4215345010332113912, -3071710691552969739, 3016784076467099616, -3592368790031611571, -8865254289946982402, -7272120709469690135, 5710438789506432976, 2727621848121380227, -6383673623215685400, 5971378783272255143, -7252475585542801376, 2661822672664531760, 4746344349502395404, 8680281578693788442, 677517871663868178, -690183649344031751, 5318163213103776061, 3934414973207186649, 5138057580394534573, -4871163826133198946, 2650305088773175077, 5611209554070066477, -4017534652661922869, 93799899709965010, 8825089704640803084, 4963003646073623934, -1711401541845829012, -4162985247158786030, 3059569130181234834, 1365639293775717036, 7384075100392926163, 4214720369944564419, -8134757447611202489, -3822250165308830692, -8699442229041762175, -8055917278538139554, 6084450476901181283, -518874445068015268, 5761099485391454117, 1508384933225121998, -5693643726318200390, 4641237854259104569, 4009157890318977485, 6784723747687817512, 5969243007480011708, 1399987878324073701, -3463542211863170711, -7449246523972958599, -815821247547853158, 9178540951995477576, -7238598586320530246, 4913926763313780375, 6378754313510757600, 3526459316021052359, 1026390510371258998, -8173384691590411473, -4939624290254720867, -4862141783623092816, -1227008064054141782, 1283053375998613198, -6322349497784455379, 2746273792355926972, -3014217871299772174, -6083762209663301754, 5628670774355814578, 899800470697144981, -1998017957699633416, 4558141621103482528, -8891165873850595166, 4866732034476510951, -643582653520169178, -7301587005367299578, -2768679825455759171, -4387564371896559615, 4818430042660156430, -6038058536635741018, 7136467645811596479, 7250937296538499324, -4770462368950287886, -6408852868267577920, -4154487122077210460, -8083736999805426, 6038134154450765143, 2336046836145680088, 1412768742431366139, -8923965341082618471, -4104653425814520335, -2653829043699544129, -3067166246055047547, -2243980054035428031, -6708369056501614003, -7129323464184410420, 2951414631009037465, 6711707486229593993, 2010179672722670783, -3673299361192059633, 6880132167760343148, 6401097090590857088, -9158727950112834240, 7800944688172644698, 1144466309505306453, -885971707798910834, 6550976832221854183, 4177986413777418037, 2350418556204242553, 3404553430623700111, 3260152572232067811, 2101138814714212542, -7364640799239757610, 1979538951335342026, 4554985479485323313, -1502882530772408662, -1862784715047786389, 1070822073551972228, 6590900980023077130, -2640964059396217697, -9138333568919549606, 1953761266339396979, 4390930362696359179] INFO [main] 2025-07-24 06:33:39,184 MigrationManager.java:331 - Create new Keyspace: KeyspaceMetadata{name=system_traces, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=2}}, tables=[org.apache.cassandra.config.CFMetaData@514718a[cfId=c5e99f16-8677-3914-b17e-960613512345,ksName=system_traces,cfName=sessions,flags=[COMPOUND],params=TableParams{comment=tracing sessions, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=0, default_time_to_live=0, memtable_flush_period_in_ms=3600000, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [client command coordinator duration request started_at parameters]],partitionKeyColumns=[session_id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UUIDType,columnMetadata=[client, command, session_id, coordinator, request, started_at, duration, parameters],droppedColumns={},triggers=[],indexes=[]], org.apache.cassandra.config.CFMetaData@479098d3[cfId=8826e8e9-e16a-3728-8753-3bc1fc713c25,ksName=system_traces,cfName=events,flags=[COMPOUND],params=TableParams{comment=tracing events, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=0, default_time_to_live=0, memtable_flush_period_in_ms=3600000, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[] | [activity source source_elapsed thread]],partitionKeyColumns=[session_id],clusteringColumns=[event_id],keyValidator=org.apache.cassandra.db.marshal.UUIDType,columnMetadata=[activity, event_id, session_id, source, thread, source_elapsed],droppedColumns={},triggers=[],indexes=[]]], views=[], functions=[], types=[]} INFO [GossipStage:1] 2025-07-24 06:33:39,567 Gossiper.java:1055 - Node /10.0.0.50 is now part of the cluster INFO [GossipStage:1] 2025-07-24 06:33:39,569 Gossiper.java:1055 - Node /10.0.0.38 is now part of the cluster INFO [RequestResponseStage-3] 2025-07-24 06:33:39,674 Gossiper.java:1019 - InetAddress /10.0.0.38 is now UP INFO [RequestResponseStage-1] 2025-07-24 06:33:39,706 Gossiper.java:1019 - InetAddress /10.0.0.50 is now UP INFO [GossipStage:1] 2025-07-24 06:33:39,728 TokenMetadata.java:479 - Updating topology for /10.0.0.38 INFO [GossipStage:1] 2025-07-24 06:33:39,728 TokenMetadata.java:479 - Updating topology for /10.0.0.38 INFO [MigrationStage:1] 2025-07-24 06:33:39,789 ViewManager.java:137 - Not submitting build tasks for views in keyspace system_traces as storage service is not initialized INFO [MigrationStage:1] 2025-07-24 06:33:39,793 ColumnFamilyStore.java:411 - Initializing system_traces.events INFO [MigrationStage:1] 2025-07-24 06:33:39,990 ColumnFamilyStore.java:411 - Initializing system_traces.sessions INFO [HANDSHAKE-/10.0.0.38] 2025-07-24 06:33:40,053 OutboundTcpConnection.java:561 - Handshaking version with /10.0.0.38 INFO [main] 2025-07-24 06:33:40,056 MigrationManager.java:331 - Create new Keyspace: KeyspaceMetadata{name=system_distributed, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=3}}, tables=[org.apache.cassandra.config.CFMetaData@113359ab[cfId=759fffad-624b-3181-80ee-fa9a52d1f627,ksName=system_distributed,cfName=repair_history,flags=[COMPOUND],params=TableParams{comment=Repair history, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=3600000, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[] | [coordinator exception_message exception_stacktrace finished_at parent_id range_begin range_end started_at status participants]],partitionKeyColumns=[keyspace_name, columnfamily_name],clusteringColumns=[id],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[status, id, coordinator, finished_at, participants, exception_stacktrace, parent_id, range_end, range_begin, exception_message, keyspace_name, started_at, columnfamily_name],droppedColumns={},triggers=[],indexes=[]], org.apache.cassandra.config.CFMetaData@581e9420[cfId=deabd734-b99d-3b9c-92e5-fd92eb5abf14,ksName=system_distributed,cfName=parent_repair_history,flags=[COMPOUND],params=TableParams{comment=Repair history, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=3600000, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [exception_message exception_stacktrace finished_at keyspace_name started_at columnfamily_names options requested_ranges successful_ranges]],partitionKeyColumns=[parent_id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[requested_ranges, exception_message, keyspace_name, successful_ranges, started_at, finished_at, options, exception_stacktrace, parent_id, columnfamily_names],droppedColumns={},triggers=[],indexes=[]], org.apache.cassandra.config.CFMetaData@577b26e5[cfId=5582b59f-8e4e-35e1-b913-3acada51eb04,ksName=system_distributed,cfName=view_build_status,flags=[COMPOUND],params=TableParams{comment=Materialized View build status, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=3600000, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.UUIDType),partitionColumns=[[] | [status]],partitionKeyColumns=[keyspace_name, view_name],clusteringColumns=[host_id],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[view_name, status, keyspace_name, host_id],droppedColumns={},triggers=[],indexes=[]]], views=[], functions=[], types=[]} INFO [HANDSHAKE-/10.0.0.50] 2025-07-24 06:33:40,098 OutboundTcpConnection.java:561 - Handshaking version with /10.0.0.50 INFO [InternalResponseStage:1] 2025-07-24 06:33:40,274 ViewManager.java:137 - Not submitting build tasks for views in keyspace system_distributed as storage service is not initialized INFO [InternalResponseStage:1] 2025-07-24 06:33:40,282 ColumnFamilyStore.java:411 - Initializing system_distributed.parent_repair_history INFO [InternalResponseStage:1] 2025-07-24 06:33:40,330 ColumnFamilyStore.java:411 - Initializing system_distributed.repair_history INFO [InternalResponseStage:1] 2025-07-24 06:33:40,385 ColumnFamilyStore.java:411 - Initializing system_distributed.view_build_status INFO [InternalResponseStage:1] 2025-07-24 06:33:40,426 ViewManager.java:137 - Not submitting build tasks for views in keyspace system_auth as storage service is not initialized INFO [InternalResponseStage:1] 2025-07-24 06:33:40,436 ColumnFamilyStore.java:411 - Initializing system_auth.resource_role_permissons_index INFO [InternalResponseStage:1] 2025-07-24 06:33:40,449 ColumnFamilyStore.java:411 - Initializing system_auth.role_members INFO [InternalResponseStage:1] 2025-07-24 06:33:40,475 ColumnFamilyStore.java:411 - Initializing system_auth.role_permissions INFO [InternalResponseStage:1] 2025-07-24 06:33:40,495 ColumnFamilyStore.java:411 - Initializing system_auth.roles INFO [main] 2025-07-24 06:33:41,419 StorageService.java:1446 - JOINING: Finish joining ring INFO [main] 2025-07-24 06:33:41,567 Gossiper.java:1692 - Waiting for gossip to settle... INFO [main] 2025-07-24 06:33:49,570 Gossiper.java:1723 - No gossip backlog; proceeding INFO [main] 2025-07-24 06:33:49,935 NativeTransportService.java:70 - Netty using native Epoll event loop INFO [main] 2025-07-24 06:33:50,031 Server.java:155 - Using Netty Version: [netty-buffer=netty-buffer-4.1.39.Final.88c2a4c (repository: dirty), netty-codec=netty-codec-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-dns=netty-codec-dns-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-haproxy=netty-codec-haproxy-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-http=netty-codec-http-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-http2=netty-codec-http2-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-memcache=netty-codec-memcache-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-mqtt=netty-codec-mqtt-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-redis=netty-codec-redis-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-smtp=netty-codec-smtp-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-socks=netty-codec-socks-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-stomp=netty-codec-stomp-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-xml=netty-codec-xml-4.1.39.Final.88c2a4c (repository: dirty), netty-common=netty-common-4.1.39.Final.88c2a4c (repository: dirty), netty-handler=netty-handler-4.1.39.Final.88c2a4c (repository: dirty), netty-handler-proxy=netty-handler-proxy-4.1.39.Final.88c2a4c (repository: dirty), netty-resolver=netty-resolver-4.1.39.Final.88c2a4c (repository: dirty), netty-resolver-dns=netty-resolver-dns-4.1.39.Final.88c2a4c (repository: dirty), netty-tcnative=netty-tcnative-2.0.25.Final.c46c351, netty-transport=netty-transport-4.1.39.Final.88c2a4c (repository: dirty), netty-transport-native-epoll=netty-transport-native-epoll-4.1.39.Final.88c2a4c (repository: dirty), netty-transport-native-kqueue=netty-transport-native-kqueue-4.1.39.Final.88c2a4cab5 (repository: dirty), netty-transport-native-unix-common=netty-transport-native-unix-common-4.1.39.Final.88c2a4c (repository: dirty), netty-transport-rxtx=netty-transport-rxtx-4.1.39.Final.88c2a4c (repository: dirty), netty-transport-sctp=netty-transport-sctp-4.1.39.Final.88c2a4c (repository: dirty), netty-transport-udt=netty-transport-udt-4.1.39.Final.88c2a4c (repository: dirty)] INFO [main] 2025-07-24 06:33:50,031 Server.java:156 - Starting listening for CQL clients on /10.0.0.254:9041 (unencrypted)... INFO [main] 2025-07-24 06:33:50,102 ThriftServer.java:116 - Binding thrift service to /10.0.0.254:9161 INFO [Thread-2] 2025-07-24 06:33:50,110 ThriftServer.java:133 - Listening for thrift clients... INFO [Native-Transport-Requests-1] 2025-07-24 06:34:00,273 MigrationManager.java:331 - Create new Keyspace: KeyspaceMetadata{name=reaper_db, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.NetworkTopologyStrategy, datacenter1=3}}, tables=[], views=[], functions=[], types=[]} INFO [Native-Transport-Requests-2] 2025-07-24 06:34:04,661 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@cb735d0[cfId=3192f250-6858-11f0-a630-edd23bd663a1,ksName=reaper_db,cfName=schema_migration,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.Int32Type),partitionColumns=[[] | [executed_at script script_name]],partitionKeyColumns=[applied_successful],clusteringColumns=[version],keyValidator=org.apache.cassandra.db.marshal.BooleanType,columnMetadata=[script_name, version, applied_successful, executed_at, script],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-07-24 06:34:04,774 ColumnFamilyStore.java:411 - Initializing reaper_db.schema_migration INFO [MigrationStage:1] 2025-07-24 06:34:05,346 ColumnFamilyStore.java:411 - Initializing reaper_db.schema_migration_leader INFO [MigrationStage:1] 2025-07-24 06:34:06,032 ColumnFamilyStore.java:411 - Initializing reaper_db.running_reapers INFO [Native-Transport-Requests-1] 2025-07-24 06:34:06,243 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@12218801[cfId=32845730-6858-11f0-a630-edd23bd663a1,ksName=reaper_db,cfName=repair_unit_v1,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [cluster_name incremental_repair keyspace_name repair_thread_count blacklisted_tables column_families datacenters nodes]],partitionKeyColumns=[id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[blacklisted_tables, datacenters, repair_thread_count, id, keyspace_name, nodes, cluster_name, incremental_repair, column_families],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-07-24 06:34:06,359 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_unit_v1 INFO [MigrationStage:1] 2025-07-24 06:34:06,711 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_schedule_by_cluster_and_keyspace INFO [Native-Transport-Requests-1] 2025-07-24 06:34:07,144 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@444cbc83[cfId=330dd280-6858-11f0-a630-edd23bd663a1,ksName=reaper_db,cfName=repair_run_by_cluster,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[] | []],partitionKeyColumns=[cluster_name],clusteringColumns=[id],keyValidator=org.apache.cassandra.db.marshal.UTF8Type,columnMetadata=[cluster_name, id],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-07-24 06:34:07,290 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_run_by_cluster INFO [MigrationStage:1] 2025-07-24 06:34:07,851 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_schedule_v1 INFO [Native-Transport-Requests-1] 2025-07-24 06:34:08,258 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@27e276b9[cfId=33b7a710-6858-11f0-a630-edd23bd663a1,ksName=reaper_db,cfName=cluster,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'ALL'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [partitioner seed_hosts]],partitionKeyColumns=[name],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UTF8Type,columnMetadata=[partitioner, seed_hosts, name],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-07-24 06:34:08,381 ColumnFamilyStore.java:411 - Initializing reaper_db.cluster INFO [MigrationStage:1] 2025-07-24 06:34:08,734 ColumnFamilyStore.java:411 - Initializing reaper_db.snapshot INFO [Native-Transport-Requests-1] 2025-07-24 06:34:09,165 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@53819e5b[cfId=344233d0-6858-11f0-a630-edd23bd663a1,ksName=reaper_db,cfName=node_metrics_v1,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=120, default_time_to_live=180, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32, unchecked_tombstone_compaction=true}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [active_anticompactions cluster datacenter has_repair_running pending_compactions requested]],partitionKeyColumns=[run_id, time_partition],clusteringColumns=[node],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UUIDType,org.apache.cassandra.db.marshal.LongType),columnMetadata=[cluster, node, has_repair_running, pending_compactions, active_anticompactions, time_partition, datacenter, requested, run_id],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-07-24 06:34:09,402 ColumnFamilyStore.java:411 - Initializing reaper_db.node_metrics_v1 INFO [MigrationStage:1] 2025-07-24 06:34:09,772 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_run INFO [Native-Transport-Requests-1] 2025-07-24 06:34:10,197 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@1bb86a07[cfId=34df8540-6858-11f0-a630-edd23bd663a1,ksName=reaper_db,cfName=repair_run_by_unit,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'ALL'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[] | []],partitionKeyColumns=[repair_unit_id],clusteringColumns=[id],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[repair_unit_id, id],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-07-24 06:34:10,354 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_run_by_unit INFO [STREAM-INIT-/10.0.0.50:53662] 2025-07-24 06:34:11,151 StreamResultFuture.java:116 - [Stream #3543ec60-6858-11f0-b5a1-4fe9959db69c ID#0] Creating new streaming plan for Bootstrap INFO [STREAM-INIT-/10.0.0.50:53662] 2025-07-24 06:34:11,163 StreamResultFuture.java:123 - [Stream #3543ec60-6858-11f0-b5a1-4fe9959db69c, ID#0] Received streaming plan for Bootstrap INFO [STREAM-INIT-/10.0.0.50:53672] 2025-07-24 06:34:11,164 StreamResultFuture.java:123 - [Stream #3543ec60-6858-11f0-b5a1-4fe9959db69c, ID#0] Received streaming plan for Bootstrap INFO [MigrationStage:1] 2025-07-24 06:34:11,401 ColumnFamilyStore.java:411 - Initializing reaper_db.leader INFO [STREAM-IN-/10.0.0.50:53672] 2025-07-24 06:34:11,409 StreamResultFuture.java:173 - [Stream #3543ec60-6858-11f0-b5a1-4fe9959db69c ID#0] Prepare completed. Receiving 0 files(0.000KiB), sending 2 files(0.173KiB) INFO [STREAM-IN-/10.0.0.50:53672] 2025-07-24 06:34:11,518 StreamResultFuture.java:187 - [Stream #3543ec60-6858-11f0-b5a1-4fe9959db69c] Session with /10.0.0.50 is complete INFO [STREAM-IN-/10.0.0.50:53672] 2025-07-24 06:34:11,519 StreamResultFuture.java:219 - [Stream #3543ec60-6858-11f0-b5a1-4fe9959db69c] All sessions completed INFO [HANDSHAKE-/10.0.0.254] 2025-07-24 06:34:12,792 OutboundTcpConnection.java:561 - Handshaking version with /10.0.0.254 WARN [Native-Transport-Requests-1] 2025-07-24 06:34:12,954 TimeFcts.java:99 - The function 'dateof' is deprecated. Use the function 'toTimestamp' instead. INFO [MigrationStage:1] 2025-07-24 06:34:14,797 ColumnFamilyStore.java:411 - Initializing reaper_db.node_metrics_v2 INFO [Native-Transport-Requests-1] 2025-07-24 06:34:15,633 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@4dbce459[cfId=381d2410-6858-11f0-a630-edd23bd663a1,ksName=reaper_db,cfName=node_operations,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=300, default_time_to_live=300, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4, unchecked_tombstone_compaction=true}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [data ts]],partitionKeyColumns=[cluster, type, time_bucket],clusteringColumns=[host],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[cluster, type, time_bucket, ts, data, host],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-07-24 06:34:15,779 ColumnFamilyStore.java:411 - Initializing reaper_db.node_operations INFO [Native-Transport-Requests-2] 2025-07-24 06:34:17,372 MigrationManager.java:454 - Update table 'reaper_db/cluster' From org.apache.cassandra.config.CFMetaData@5e771b72[cfId=33b7a710-6858-11f0-a630-edd23bd663a1,ksName=reaper_db,cfName=cluster,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'ALL'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [partitioner properties seed_hosts]],partitionKeyColumns=[name],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UTF8Type,columnMetadata=[properties, name, partitioner, seed_hosts],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@1cd14d43[cfId=33b7a710-6858-11f0-a630-edd23bd663a1,ksName=reaper_db,cfName=cluster,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'ALL'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [partitioner properties state seed_hosts]],partitionKeyColumns=[name],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UTF8Type,columnMetadata=[properties, state, name, partitioner, seed_hosts],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-07-24 06:34:18,907 ColumnFamilyStore.java:411 - Initializing reaper_db.diagnostic_event_subscription INFO [Native-Transport-Requests-1] 2025-07-24 06:34:20,632 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@793f1074[cfId=3b17ed80-6858-11f0-a630-edd23bd663a1,ksName=reaper_db,cfName=node_metrics_v3,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=300, default_time_to_live=3600, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4, unchecked_tombstone_compaction=true}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.ReversedType(org.apache.cassandra.db.marshal.TimestampType), org.apache.cassandra.db.marshal.UTF8Type, org.apache.cassandra.db.marshal.UTF8Type, org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [value]],partitionKeyColumns=[cluster, metric_domain, metric_type, time_bucket, host],clusteringColumns=[ts, metric_scope, metric_name, metric_attribute],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[cluster, metric_domain, metric_attribute, time_bucket, ts, metric_type, metric_name, metric_scope, value, host],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-07-24 06:34:20,777 ColumnFamilyStore.java:411 - Initializing reaper_db.node_metrics_v3 INFO [Native-Transport-Requests-2] 2025-07-24 06:34:21,305 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@389ca54e[cfId=3b7e9e90-6858-11f0-a630-edd23bd663a1,ksName=reaper_db,cfName=repair_run_by_cluster_v2,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.ReversedType(org.apache.cassandra.db.marshal.TimeUUIDType)),partitionColumns=[[] | [repair_run_state]],partitionKeyColumns=[cluster_name],clusteringColumns=[id],keyValidator=org.apache.cassandra.db.marshal.UTF8Type,columnMetadata=[cluster_name, repair_run_state, id],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-07-24 06:34:21,441 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_run_by_cluster_v2 INFO [Native-Transport-Requests-1] 2025-07-24 06:34:21,780 MigrationManager.java:454 - Update table 'reaper_db/repair_run' From org.apache.cassandra.config.CFMetaData@261675de[cfId=348afca0-6858-11f0-bd5d-ef813b7a18da,ksName=reaper_db,cfName=repair_run,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : '5000'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[cause cluster_name creation_time end_time intensity last_event owner pause_time repair_parallelism repair_unit_id segment_count start_time state tables] | [coordinator_host end_token fail_count segment_end_time segment_start_time segment_state start_token token_ranges]],partitionKeyColumns=[id],clusteringColumns=[segment_id],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[intensity, segment_count, last_event, id, segment_end_time, state, cluster_name, end_time, end_token, start_token, segment_start_time, segment_state, cause, creation_time, start_time, coordinator_host, token_ranges, owner, repair_parallelism, tables, segment_id, pause_time, repair_unit_id, fail_count],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@1cdfc22[cfId=348afca0-6858-11f0-bd5d-ef813b7a18da,ksName=reaper_db,cfName=repair_run,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : '5000'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[cause cluster_name creation_time end_time intensity last_event owner pause_time repair_parallelism repair_unit_id segment_count start_time state tables] | [coordinator_host end_token fail_count replicas segment_end_time segment_start_time segment_state start_token token_ranges]],partitionKeyColumns=[id],clusteringColumns=[segment_id],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[intensity, id, segment_end_time, state, end_token, start_token, start_time, token_ranges, tables, pause_time, repair_unit_id, segment_count, last_event, cluster_name, end_time, segment_start_time, segment_state, cause, creation_time, coordinator_host, replicas, owner, repair_parallelism, segment_id, fail_count],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-07-24 06:34:22,898 ColumnFamilyStore.java:411 - Initializing reaper_db.running_repairs INFO [MigrationStage:1] 2025-07-24 06:34:23,789 ColumnFamilyStore.java:411 - Initializing reaper_db.percent_repaired_by_schedule INFO [Native-Transport-Requests-2] 2025-07-24 06:34:24,683 MigrationManager.java:454 - Update table 'reaper_db/repair_unit_v1' From org.apache.cassandra.config.CFMetaData@4477eb7a[cfId=32845730-6858-11f0-a630-edd23bd663a1,ksName=reaper_db,cfName=repair_unit_v1,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [cluster_name incremental_repair keyspace_name repair_thread_count blacklisted_tables column_families datacenters nodes]],partitionKeyColumns=[id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[blacklisted_tables, datacenters, repair_thread_count, id, keyspace_name, nodes, cluster_name, incremental_repair, column_families],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@4ea2371a[cfId=32845730-6858-11f0-a630-edd23bd663a1,ksName=reaper_db,cfName=repair_unit_v1,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [cluster_name incremental_repair keyspace_name repair_thread_count timeout blacklisted_tables column_families datacenters nodes]],partitionKeyColumns=[id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[blacklisted_tables, datacenters, repair_thread_count, id, keyspace_name, timeout, nodes, cluster_name, incremental_repair, column_families],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-3] 2025-07-24 06:34:27,863 MigrationManager.java:454 - Update table 'reaper_db/repair_schedule_v1' From org.apache.cassandra.config.CFMetaData@1632066f[cfId=336604a0-6858-11f0-bd5d-ef813b7a18da,ksName=reaper_db,cfName=repair_schedule_v1,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [adaptive creation_time days_between intensity next_activation owner pause_time repair_parallelism repair_unit_id segment_count segment_count_per_node state run_history]],partitionKeyColumns=[id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[intensity, segment_count, days_between, id, state, run_history, creation_time, adaptive, owner, repair_parallelism, segment_count_per_node, pause_time, repair_unit_id, next_activation],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@64f53efb[cfId=336604a0-6858-11f0-bd5d-ef813b7a18da,ksName=reaper_db,cfName=repair_schedule_v1,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [adaptive creation_time days_between intensity next_activation owner pause_time percent_unrepaired_threshold repair_parallelism repair_unit_id segment_count segment_count_per_node state run_history]],partitionKeyColumns=[id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[intensity, segment_count, days_between, percent_unrepaired_threshold, id, state, run_history, creation_time, adaptive, owner, repair_parallelism, segment_count_per_node, pause_time, repair_unit_id, next_activation],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-2] 2025-07-24 06:34:30,661 MigrationManager.java:454 - Update table 'reaper_db/diagnostic_event_subscription' From org.apache.cassandra.config.CFMetaData@59cc8473[cfId=39f9d350-6858-11f0-bd5d-ef813b7a18da,ksName=reaper_db,cfName=diagnostic_event_subscription,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'ALL'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [cluster description export_file_logger export_http_endpoint export_sse events nodes]],partitionKeyColumns=[id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UUIDType,columnMetadata=[cluster, export_http_endpoint, events, id, export_sse, nodes, export_file_logger, description],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@7f33b39c[cfId=39f9d350-6858-11f0-bd5d-ef813b7a18da,ksName=reaper_db,cfName=diagnostic_event_subscription,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'ALL'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [cluster description export_file_logger export_http_endpoint export_sse events nodes]],partitionKeyColumns=[id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UUIDType,columnMetadata=[cluster, export_http_endpoint, events, id, export_sse, nodes, export_file_logger, description],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-1] 2025-07-24 06:34:30,663 MigrationManager.java:454 - Update table 'reaper_db/node_metrics_v3' From org.apache.cassandra.config.CFMetaData@e8bdc06[cfId=3b17ed80-6858-11f0-a630-edd23bd663a1,ksName=reaper_db,cfName=node_metrics_v3,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=300, default_time_to_live=3600, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4, unchecked_tombstone_compaction=true}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.ReversedType(org.apache.cassandra.db.marshal.TimestampType), org.apache.cassandra.db.marshal.UTF8Type, org.apache.cassandra.db.marshal.UTF8Type, org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [value]],partitionKeyColumns=[cluster, metric_domain, metric_type, time_bucket, host],clusteringColumns=[ts, metric_scope, metric_name, metric_attribute],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[cluster, metric_domain, metric_attribute, time_bucket, ts, metric_type, metric_name, metric_scope, value, host],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@37947101[cfId=3b17ed80-6858-11f0-a630-edd23bd663a1,ksName=reaper_db,cfName=node_metrics_v3,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=300, default_time_to_live=3600, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4, unchecked_tombstone_compaction=true}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.ReversedType(org.apache.cassandra.db.marshal.TimestampType), org.apache.cassandra.db.marshal.UTF8Type, org.apache.cassandra.db.marshal.UTF8Type, org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [value]],partitionKeyColumns=[cluster, metric_domain, metric_type, time_bucket, host],clusteringColumns=[ts, metric_scope, metric_name, metric_attribute],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[cluster, metric_domain, metric_attribute, time_bucket, ts, metric_type, metric_name, metric_scope, value, host],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-3] 2025-07-24 06:34:30,667 MigrationManager.java:454 - Update table 'reaper_db/repair_run' From org.apache.cassandra.config.CFMetaData@261675de[cfId=348afca0-6858-11f0-bd5d-ef813b7a18da,ksName=reaper_db,cfName=repair_run,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : '5000'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[adaptive_schedule cause cluster_name creation_time end_time intensity last_event owner pause_time repair_parallelism repair_unit_id segment_count start_time state tables] | [coordinator_host end_token fail_count host_id replicas segment_end_time segment_start_time segment_state start_token token_ranges]],partitionKeyColumns=[id],clusteringColumns=[segment_id],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[intensity, id, segment_end_time, state, end_token, start_token, start_time, token_ranges, tables, pause_time, repair_unit_id, host_id, segment_count, last_event, adaptive_schedule, cluster_name, end_time, segment_start_time, segment_state, cause, creation_time, coordinator_host, replicas, owner, repair_parallelism, segment_id, fail_count],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@46496f47[cfId=348afca0-6858-11f0-bd5d-ef813b7a18da,ksName=reaper_db,cfName=repair_run,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : '5000'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[adaptive_schedule cause cluster_name creation_time end_time intensity last_event owner pause_time repair_parallelism repair_unit_id segment_count start_time state tables] | [coordinator_host end_token fail_count host_id replicas segment_end_time segment_start_time segment_state start_token token_ranges]],partitionKeyColumns=[id],clusteringColumns=[segment_id],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[intensity, id, segment_end_time, state, end_token, start_token, start_time, token_ranges, tables, pause_time, repair_unit_id, host_id, segment_count, last_event, adaptive_schedule, cluster_name, end_time, segment_start_time, segment_state, cause, creation_time, coordinator_host, replicas, owner, repair_parallelism, segment_id, fail_count],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-4] 2025-07-24 06:34:30,669 MigrationManager.java:454 - Update table 'reaper_db/running_reapers' From org.apache.cassandra.config.CFMetaData@607ec97b[cfId=324f8b90-6858-11f0-bd5d-ef813b7a18da,ksName=reaper_db,cfName=running_reapers,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=180, default_time_to_live=180, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [last_heartbeat reaper_instance_host]],partitionKeyColumns=[reaper_instance_id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UUIDType,columnMetadata=[reaper_instance_id, last_heartbeat, reaper_instance_host],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@3ecbaa46[cfId=324f8b90-6858-11f0-bd5d-ef813b7a18da,ksName=reaper_db,cfName=running_reapers,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=180, default_time_to_live=180, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [last_heartbeat reaper_instance_host]],partitionKeyColumns=[reaper_instance_id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UUIDType,columnMetadata=[reaper_instance_id, last_heartbeat, reaper_instance_host],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-5] 2025-07-24 06:34:30,670 MigrationManager.java:454 - Update table 'reaper_db/repair_run_by_unit' From org.apache.cassandra.config.CFMetaData@7a721b41[cfId=34df8540-6858-11f0-a630-edd23bd663a1,ksName=reaper_db,cfName=repair_run_by_unit,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'ALL'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[] | []],partitionKeyColumns=[repair_unit_id],clusteringColumns=[id],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[repair_unit_id, id],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@5d970068[cfId=34df8540-6858-11f0-a630-edd23bd663a1,ksName=reaper_db,cfName=repair_run_by_unit,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'ALL'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[] | []],partitionKeyColumns=[repair_unit_id],clusteringColumns=[id],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[repair_unit_id, id],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-6] 2025-07-24 06:34:30,672 MigrationManager.java:454 - Update table 'reaper_db/schema_migration' From org.apache.cassandra.config.CFMetaData@1dd7cd9f[cfId=3192f250-6858-11f0-a630-edd23bd663a1,ksName=reaper_db,cfName=schema_migration,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.Int32Type),partitionColumns=[[] | [executed_at script script_name]],partitionKeyColumns=[applied_successful],clusteringColumns=[version],keyValidator=org.apache.cassandra.db.marshal.BooleanType,columnMetadata=[script_name, version, applied_successful, executed_at, script],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@50b411c4[cfId=3192f250-6858-11f0-a630-edd23bd663a1,ksName=reaper_db,cfName=schema_migration,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.Int32Type),partitionColumns=[[] | [executed_at script script_name]],partitionKeyColumns=[applied_successful],clusteringColumns=[version],keyValidator=org.apache.cassandra.db.marshal.BooleanType,columnMetadata=[script_name, version, applied_successful, executed_at, script],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-3] 2025-07-24 06:34:35,747 MigrationManager.java:454 - Update table 'reaper_db/node_operations' From org.apache.cassandra.config.CFMetaData@24228bc[cfId=381d2410-6858-11f0-a630-edd23bd663a1,ksName=reaper_db,cfName=node_operations,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=300, default_time_to_live=300, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4, unchecked_tombstone_compaction=true}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [data ts]],partitionKeyColumns=[cluster, type, time_bucket],clusteringColumns=[host],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[cluster, type, time_bucket, ts, data, host],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@7ff1b0bd[cfId=381d2410-6858-11f0-a630-edd23bd663a1,ksName=reaper_db,cfName=node_operations,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=300, default_time_to_live=300, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.TimeWindowCompactionStrategy, options={min_threshold=4, max_threshold=32, compaction_window_size=30, compaction_window_unit=MINUTES, unchecked_tombstone_compaction=true}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [data ts]],partitionKeyColumns=[cluster, type, time_bucket],clusteringColumns=[host],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[cluster, type, time_bucket, ts, data, host],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-1] 2025-07-24 06:34:37,227 MigrationManager.java:427 - Update Keyspace 'svc_monitor_keyspace' From KeyspaceMetadata{name=svc_monitor_keyspace, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=3}}, tables=[], views=[], functions=[], types=[]} To KeyspaceMetadata{name=svc_monitor_keyspace, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=3}}, tables=[], views=[], functions=[], types=[]} INFO [MigrationStage:1] 2025-07-24 06:34:38,325 ColumnFamilyStore.java:411 - Initializing config_db_uuid.obj_uuid_table INFO [MigrationStage:1] 2025-07-24 06:34:39,878 ColumnFamilyStore.java:411 - Initializing svc_monitor_keyspace.service_instance_table INFO [Native-Transport-Requests-1] 2025-07-24 06:34:40,782 MigrationManager.java:454 - Update table 'config_db_uuid/obj_uuid_table' From org.apache.cassandra.config.CFMetaData@7597068f[cfId=4591a8f0-6858-11f0-b5a1-4fe9959db69c,ksName=config_db_uuid,cfName=obj_uuid_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@15136417[cfId=4591a8f0-6858-11f0-b5a1-4fe9959db69c,ksName=config_db_uuid,cfName=obj_uuid_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-4] 2025-07-24 06:34:41,521 MigrationManager.java:331 - Create new Keyspace: KeyspaceMetadata{name=useragent, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=3}}, tables=[], views=[], functions=[], types=[]} INFO [MigrationStage:1] 2025-07-24 06:34:43,636 ColumnFamilyStore.java:411 - Initializing config_db_uuid.obj_fq_name_table INFO [Native-Transport-Requests-1] 2025-07-24 06:34:45,774 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@69d2e5e[cfId=4a144ae0-6858-11f0-a630-edd23bd663a1,ksName=svc_monitor_keyspace,cfName=pool_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-07-24 06:34:45,891 ColumnFamilyStore.java:411 - Initializing svc_monitor_keyspace.pool_table INFO [MigrationStage:1] 2025-07-24 06:34:48,791 ColumnFamilyStore.java:411 - Initializing config_db_uuid.obj_shared_table INFO [MigrationStage:1] 2025-07-24 06:34:52,281 ColumnFamilyStore.java:411 - Initializing svc_monitor_keyspace.loadbalancer_table INFO [Native-Transport-Requests-1] 2025-07-24 06:34:54,728 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@627f7bb2[cfId=4f6a9080-6858-11f0-a630-edd23bd663a1,ksName=useragent,cfName=useragent_keyval_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-07-24 06:34:54,833 ColumnFamilyStore.java:411 - Initializing useragent.useragent_keyval_table INFO [Native-Transport-Requests-4] 2025-07-24 06:34:55,677 MigrationManager.java:454 - Update table 'svc_monitor_keyspace/loadbalancer_table' From org.apache.cassandra.config.CFMetaData@3d852dd5[cfId=4de59e30-6858-11f0-b5a1-4fe9959db69c,ksName=svc_monitor_keyspace,cfName=loadbalancer_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@7caab8fd[cfId=4de59e30-6858-11f0-b5a1-4fe9959db69c,ksName=svc_monitor_keyspace,cfName=loadbalancer_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-07-24 06:34:57,688 ColumnFamilyStore.java:411 - Initializing svc_monitor_keyspace.healthmonitor_table INFO [Native-Transport-Requests-1] 2025-07-24 06:35:00,581 MigrationManager.java:454 - Update table 'config_db_uuid/obj_shared_table' From org.apache.cassandra.config.CFMetaData@1d40825e[cfId=4bce56f0-6858-11f0-b5a1-4fe9959db69c,ksName=config_db_uuid,cfName=obj_shared_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@60a07e50[cfId=4bce56f0-6858-11f0-b5a1-4fe9959db69c,ksName=config_db_uuid,cfName=obj_shared_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-07-24 06:35:03,401 ColumnFamilyStore.java:411 - Initializing to_bgp_keyspace.route_target_table INFO [Native-Transport-Requests-2] 2025-07-24 06:35:03,623 MigrationManager.java:454 - Update table 'to_bgp_keyspace/route_target_table' From org.apache.cassandra.config.CFMetaData@6d6f0710[cfId=54841b40-6858-11f0-bd5d-ef813b7a18da,ksName=to_bgp_keyspace,cfName=route_target_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@3074e8af[cfId=54841b40-6858-11f0-bd5d-ef813b7a18da,ksName=to_bgp_keyspace,cfName=route_target_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-07-24 06:35:04,997 ColumnFamilyStore.java:411 - Initializing to_bgp_keyspace.service_chain_ip_address_table INFO [Native-Transport-Requests-4] 2025-07-24 06:35:06,706 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@6d9f942d[cfId=568e4320-6858-11f0-a630-edd23bd663a1,ksName=to_bgp_keyspace,cfName=service_chain_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-07-24 06:35:06,921 ColumnFamilyStore.java:411 - Initializing to_bgp_keyspace.service_chain_table INFO [MigrationStage:1] 2025-07-24 06:35:08,835 ColumnFamilyStore.java:411 - Initializing to_bgp_keyspace.service_chain_uuid_table INFO [Native-Transport-Requests-1] 2025-07-24 06:35:09,580 MigrationManager.java:454 - Update table 'to_bgp_keyspace/service_chain_uuid_table' From org.apache.cassandra.config.CFMetaData@dd00155[cfId=57b1db90-6858-11f0-bd5d-ef813b7a18da,ksName=to_bgp_keyspace,cfName=service_chain_uuid_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@6c4c29e4[cfId=57b1db90-6858-11f0-bd5d-ef813b7a18da,ksName=to_bgp_keyspace,cfName=service_chain_uuid_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-2] 2025-07-24 06:36:26,854 MigrationManager.java:427 - Update Keyspace 'dm_keyspace' From KeyspaceMetadata{name=dm_keyspace, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=3}}, tables=[], views=[], functions=[], types=[]} To KeyspaceMetadata{name=dm_keyspace, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=3}}, tables=[], views=[], functions=[], types=[]} INFO [MigrationStage:1] 2025-07-24 06:36:28,138 ColumnFamilyStore.java:411 - Initializing dm_keyspace.dm_pr_vn_ip_table INFO [Native-Transport-Requests-1] 2025-07-24 06:36:28,851 MigrationManager.java:454 - Update table 'dm_keyspace/dm_pr_vn_ip_table' From org.apache.cassandra.config.CFMetaData@267d6847[cfId=86ef8420-6858-11f0-b5a1-4fe9959db69c,ksName=dm_keyspace,cfName=dm_pr_vn_ip_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@7936959f[cfId=86ef8420-6858-11f0-b5a1-4fe9959db69c,ksName=dm_keyspace,cfName=dm_pr_vn_ip_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-07-24 06:36:29,968 ColumnFamilyStore.java:411 - Initializing dm_keyspace.dm_pr_asn_table INFO [Native-Transport-Requests-1] 2025-07-24 06:36:31,730 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@68422814[cfId=893bbe10-6858-11f0-a630-edd23bd663a1,ksName=dm_keyspace,cfName=dm_ni_ipv6_ll_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-07-24 06:36:31,824 ColumnFamilyStore.java:411 - Initializing dm_keyspace.dm_ni_ipv6_ll_table INFO [MigrationStage:1] 2025-07-24 06:36:34,065 ColumnFamilyStore.java:411 - Initializing dm_keyspace.dm_pnf_resource_table INFO [Native-Transport-Requests-1] 2025-07-24 06:36:34,665 MigrationManager.java:454 - Update table 'dm_keyspace/dm_pnf_resource_table' From org.apache.cassandra.config.CFMetaData@551f6e24[cfId=8a7ec560-6858-11f0-b5a1-4fe9959db69c,ksName=dm_keyspace,cfName=dm_pnf_resource_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@7adca498[cfId=8a7ec560-6858-11f0-b5a1-4fe9959db69c,ksName=dm_keyspace,cfName=dm_pnf_resource_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@2c987619, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [HANDSHAKE-/10.0.0.254] 2025-07-24 06:41:43,023 OutboundTcpConnection.java:561 - Handshaking version with /10.0.0.254 INFO [Repair-Task-2] 2025-07-24 06:41:43,367 RepairRunnable.java:139 - Starting repair command #1 (42fbed70-6859-11f0-a630-edd23bd663a1), repairing keyspace reaper_db with repair options (parallelism: dc_parallel, primary range: false, incremental: false, job threads: 1, ColumnFamilies: [cluster, leader, diagnostic_event_subscription, schema_migration_leader, running_repairs, repair_run, repair_unit_v1, repair_run_by_cluster_v2, repair_schedule_by_cluster_and_keyspace, repair_run_by_cluster, running_reapers, repair_schedule_v1, percent_repaired_by_schedule, repair_run_by_unit, schema_migration, snapshot], dataCenters: [], hosts: [], # of ranges: 1, pull repair: false) INFO [Repair-Task-2] 2025-07-24 06:41:43,438 RepairSession.java:228 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] new session: will sync /10.0.0.254, /10.0.0.50, /10.0.0.38 on range [(-1711401541845829012,-1624776244370520620]] for reaper_db.[percent_repaired_by_schedule, repair_run_by_cluster_v2, repair_schedule_v1, snapshot, repair_run_by_cluster, repair_run_by_unit, cluster, schema_migration_leader, leader, diagnostic_event_subscription, running_repairs, repair_run, repair_unit_v1, repair_schedule_by_cluster_and_keyspace, schema_migration, running_reapers] INFO [RepairJobTask:3] 2025-07-24 06:41:43,498 RepairJob.java:234 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for percent_repaired_by_schedule (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:3] 2025-07-24 06:41:43,500 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,533 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for percent_repaired_by_schedule from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,534 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,555 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for percent_repaired_by_schedule from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,556 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,568 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for percent_repaired_by_schedule from /10.0.0.254 INFO [RepairJobTask:1] 2025-07-24 06:41:43,570 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:3] 2025-07-24 06:41:43,571 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:5] 2025-07-24 06:41:43,571 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:3] 2025-07-24 06:41:43,571 RepairJob.java:143 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] percent_repaired_by_schedule is fully synced INFO [RepairJobTask:3] 2025-07-24 06:41:43,623 RepairJob.java:234 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_run_by_cluster_v2 (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:3] 2025-07-24 06:41:43,623 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,625 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,626 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,628 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,629 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,631 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.254 INFO [RepairJobTask:3] 2025-07-24 06:41:43,632 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:3] 2025-07-24 06:41:43,633 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:3] 2025-07-24 06:41:43,633 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:3] 2025-07-24 06:41:43,633 RepairJob.java:143 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] repair_run_by_cluster_v2 is fully synced INFO [RepairJobTask:3] 2025-07-24 06:41:43,690 RepairJob.java:234 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_schedule_v1 (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:3] 2025-07-24 06:41:43,690 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,695 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_v1 from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,695 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,702 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_v1 from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,702 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,708 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_v1 from /10.0.0.254 INFO [RepairJobTask:4] 2025-07-24 06:41:43,709 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_schedule_v1 INFO [RepairJobTask:2] 2025-07-24 06:41:43,709 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_schedule_v1 INFO [RepairJobTask:5] 2025-07-24 06:41:43,710 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_schedule_v1 INFO [RepairJobTask:2] 2025-07-24 06:41:43,710 RepairJob.java:143 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] repair_schedule_v1 is fully synced INFO [RepairJobTask:2] 2025-07-24 06:41:43,718 RepairJob.java:234 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for snapshot (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:2] 2025-07-24 06:41:43,718 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,722 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for snapshot from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,723 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,725 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for snapshot from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,725 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,730 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for snapshot from /10.0.0.254 INFO [RepairJobTask:4] 2025-07-24 06:41:43,731 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for snapshot INFO [RepairJobTask:4] 2025-07-24 06:41:43,731 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for snapshot INFO [RepairJobTask:2] 2025-07-24 06:41:43,731 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for snapshot INFO [RepairJobTask:4] 2025-07-24 06:41:43,732 RepairJob.java:143 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] snapshot is fully synced INFO [RepairJobTask:4] 2025-07-24 06:41:43,736 RepairJob.java:234 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_run_by_cluster (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:4] 2025-07-24 06:41:43,737 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,740 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,740 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,742 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,742 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,743 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster from /10.0.0.254 INFO [RepairJobTask:3] 2025-07-24 06:41:43,744 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_run_by_cluster INFO [RepairJobTask:4] 2025-07-24 06:41:43,744 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_run_by_cluster INFO [RepairJobTask:2] 2025-07-24 06:41:43,744 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_run_by_cluster INFO [RepairJobTask:4] 2025-07-24 06:41:43,744 RepairJob.java:143 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] repair_run_by_cluster is fully synced INFO [RepairJobTask:4] 2025-07-24 06:41:43,798 RepairJob.java:234 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_run_by_unit (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:4] 2025-07-24 06:41:43,799 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,800 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_unit from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,801 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,803 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_unit from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,804 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,806 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_unit from /10.0.0.254 INFO [RepairJobTask:1] 2025-07-24 06:41:43,806 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_run_by_unit INFO [RepairJobTask:5] 2025-07-24 06:41:43,806 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_run_by_unit INFO [RepairJobTask:3] 2025-07-24 06:41:43,806 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_run_by_unit INFO [RepairJobTask:5] 2025-07-24 06:41:43,807 RepairJob.java:143 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] repair_run_by_unit is fully synced INFO [RepairJobTask:5] 2025-07-24 06:41:43,850 RepairJob.java:234 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for cluster (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:5] 2025-07-24 06:41:43,851 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,853 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for cluster from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,854 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,857 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for cluster from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,857 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,858 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for cluster from /10.0.0.254 INFO [RepairJobTask:5] 2025-07-24 06:41:43,858 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for cluster INFO [RepairJobTask:3] 2025-07-24 06:41:43,858 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for cluster INFO [RepairJobTask:2] 2025-07-24 06:41:43,859 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for cluster INFO [RepairJobTask:1] 2025-07-24 06:41:43,859 RepairJob.java:143 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] cluster is fully synced INFO [RepairJobTask:6] 2025-07-24 06:41:43,879 RepairJob.java:234 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for schema_migration_leader (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:6] 2025-07-24 06:41:43,879 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,882 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration_leader from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,882 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,884 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration_leader from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,884 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,887 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration_leader from /10.0.0.254 INFO [RepairJobTask:2] 2025-07-24 06:41:43,888 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for schema_migration_leader INFO [RepairJobTask:3] 2025-07-24 06:41:43,888 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for schema_migration_leader INFO [RepairJobTask:6] 2025-07-24 06:41:43,888 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for schema_migration_leader INFO [RepairJobTask:1] 2025-07-24 06:41:43,888 RepairJob.java:143 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] schema_migration_leader is fully synced INFO [RepairJobTask:1] 2025-07-24 06:41:43,890 RepairJob.java:234 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for leader (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:1] 2025-07-24 06:41:43,890 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,895 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for leader from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,895 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,897 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for leader from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,897 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,899 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for leader from /10.0.0.254 INFO [RepairJobTask:2] 2025-07-24 06:41:43,899 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for leader INFO [RepairJobTask:5] 2025-07-24 06:41:43,899 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for leader INFO [RepairJobTask:3] 2025-07-24 06:41:43,899 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for leader INFO [RepairJobTask:5] 2025-07-24 06:41:43,900 RepairJob.java:143 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] leader is fully synced INFO [RepairJobTask:6] 2025-07-24 06:41:43,904 RepairJob.java:234 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for diagnostic_event_subscription (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:6] 2025-07-24 06:41:43,905 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,908 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for diagnostic_event_subscription from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,908 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,911 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for diagnostic_event_subscription from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,911 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,913 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for diagnostic_event_subscription from /10.0.0.254 INFO [RepairJobTask:5] 2025-07-24 06:41:43,913 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for diagnostic_event_subscription INFO [RepairJobTask:5] 2025-07-24 06:41:43,913 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for diagnostic_event_subscription INFO [RepairJobTask:5] 2025-07-24 06:41:43,913 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for diagnostic_event_subscription INFO [RepairJobTask:6] 2025-07-24 06:41:43,914 RepairJob.java:143 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] diagnostic_event_subscription is fully synced INFO [RepairJobTask:6] 2025-07-24 06:41:43,963 RepairJob.java:234 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for running_repairs (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:6] 2025-07-24 06:41:43,963 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,966 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_repairs from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,966 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,969 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_repairs from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,969 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:43,970 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_repairs from /10.0.0.254 INFO [RepairJobTask:5] 2025-07-24 06:41:43,971 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for running_repairs INFO [RepairJobTask:2] 2025-07-24 06:41:43,971 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for running_repairs INFO [RepairJobTask:5] 2025-07-24 06:41:43,971 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for running_repairs INFO [RepairJobTask:5] 2025-07-24 06:41:43,971 RepairJob.java:143 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] running_repairs is fully synced INFO [RepairJobTask:5] 2025-07-24 06:41:44,023 RepairJob.java:234 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_run (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:5] 2025-07-24 06:41:44,023 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:44,025 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:44,026 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:44,035 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:44,035 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:44,038 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run from /10.0.0.254 INFO [RepairJobTask:3] 2025-07-24 06:41:44,039 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_run INFO [RepairJobTask:1] 2025-07-24 06:41:44,039 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_run INFO [RepairJobTask:2] 2025-07-24 06:41:44,039 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_run INFO [RepairJobTask:1] 2025-07-24 06:41:44,039 RepairJob.java:143 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] repair_run is fully synced INFO [RepairJobTask:1] 2025-07-24 06:41:44,082 RepairJob.java:234 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_unit_v1 (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:1] 2025-07-24 06:41:44,082 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:44,084 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_unit_v1 from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:44,085 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:44,090 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_unit_v1 from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:44,090 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:44,092 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_unit_v1 from /10.0.0.254 INFO [RepairJobTask:6] 2025-07-24 06:41:44,092 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_unit_v1 INFO [RepairJobTask:4] 2025-07-24 06:41:44,092 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_unit_v1 INFO [RepairJobTask:7] 2025-07-24 06:41:44,093 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_unit_v1 INFO [RepairJobTask:5] 2025-07-24 06:41:44,093 RepairJob.java:143 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] repair_unit_v1 is fully synced INFO [RepairJobTask:5] 2025-07-24 06:41:44,146 RepairJob.java:234 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_schedule_by_cluster_and_keyspace (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:5] 2025-07-24 06:41:44,146 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:44,148 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:44,148 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:44,156 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:44,156 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:44,159 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.254 INFO [RepairJobTask:6] 2025-07-24 06:41:44,160 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:3] 2025-07-24 06:41:44,160 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:2] 2025-07-24 06:41:44,160 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:3] 2025-07-24 06:41:44,160 RepairJob.java:143 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] repair_schedule_by_cluster_and_keyspace is fully synced INFO [RepairJobTask:3] 2025-07-24 06:41:44,163 RepairJob.java:234 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for schema_migration (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:3] 2025-07-24 06:41:44,163 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:44,165 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:44,166 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:44,172 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:44,172 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:44,173 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration from /10.0.0.254 INFO [RepairJobTask:6] 2025-07-24 06:41:44,174 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for schema_migration INFO [RepairJobTask:2] 2025-07-24 06:41:44,174 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for schema_migration INFO [RepairJobTask:5] 2025-07-24 06:41:44,174 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for schema_migration INFO [RepairJobTask:2] 2025-07-24 06:41:44,174 RepairJob.java:143 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] schema_migration is fully synced INFO [RepairJobTask:2] 2025-07-24 06:41:44,225 RepairJob.java:234 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for running_reapers (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:2] 2025-07-24 06:41:44,225 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:44,227 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_reapers from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:44,227 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:44,230 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_reapers from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:44,230 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:44,232 RepairSession.java:180 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_reapers from /10.0.0.254 INFO [RepairJobTask:2] 2025-07-24 06:41:44,233 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for running_reapers INFO [RepairJobTask:6] 2025-07-24 06:41:44,233 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for running_reapers INFO [RepairJobTask:3] 2025-07-24 06:41:44,233 SyncTask.java:66 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for running_reapers INFO [RepairJobTask:6] 2025-07-24 06:41:44,233 RepairJob.java:143 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] running_reapers is fully synced INFO [RepairJobTask:6] 2025-07-24 06:41:44,235 RepairSession.java:270 - [repair #43064db0-6859-11f0-a630-edd23bd663a1] Session completed successfully INFO [RepairJobTask:6] 2025-07-24 06:41:44,236 RepairRunnable.java:261 - Repair session 43064db0-6859-11f0-a630-edd23bd663a1 for range [(-1711401541845829012,-1624776244370520620]] finished INFO [RepairJobTask:6] 2025-07-24 06:41:44,237 ActiveRepairService.java:452 - [repair #42fbed70-6859-11f0-a630-edd23bd663a1] Not a global repair, will not do anticompaction INFO [InternalResponseStage:5] 2025-07-24 06:41:44,245 RepairRunnable.java:343 - Repair command #1 finished in 0 seconds INFO [AntiEntropyStage:1] 2025-07-24 06:41:52,818 Validator.java:281 - [repair #488996c0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.diagnostic_event_subscription INFO [AntiEntropyStage:1] 2025-07-24 06:41:52,839 Validator.java:281 - [repair #488996c0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.percent_repaired_by_schedule INFO [AntiEntropyStage:1] 2025-07-24 06:41:52,900 Validator.java:281 - [repair #488996c0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.repair_run INFO [AntiEntropyStage:1] 2025-07-24 06:41:52,919 Validator.java:281 - [repair #488996c0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.schema_migration_leader INFO [AntiEntropyStage:1] 2025-07-24 06:41:52,945 Validator.java:281 - [repair #488996c0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.repair_run_by_unit INFO [AntiEntropyStage:1] 2025-07-24 06:41:52,964 Validator.java:281 - [repair #488996c0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.running_reapers INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,025 Validator.java:281 - [repair #488996c0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.running_repairs INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,043 Validator.java:281 - [repair #488996c0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.repair_schedule_v1 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,057 Validator.java:281 - [repair #488996c0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.cluster INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,074 Validator.java:281 - [repair #488996c0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.repair_unit_v1 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,092 Validator.java:281 - [repair #488996c0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.repair_run_by_cluster INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,103 Validator.java:281 - [repair #488996c0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.snapshot INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,120 Validator.java:281 - [repair #488996c0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.leader INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,132 Validator.java:281 - [repair #488996c0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.repair_run_by_cluster_v2 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,147 Validator.java:281 - [repair #488996c0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.repair_schedule_by_cluster_and_keyspace INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,162 Validator.java:281 - [repair #488996c0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.schema_migration INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,174 ActiveRepairService.java:452 - [repair #48843f90-6859-11f0-bd5d-ef813b7a18da] Not a global repair, will not do anticompaction INFO [HANDSHAKE-/10.0.0.38] 2025-07-24 06:41:53,331 OutboundTcpConnection.java:561 - Handshaking version with /10.0.0.38 INFO [Repair-Task-3] 2025-07-24 06:41:53,481 RepairRunnable.java:139 - Starting repair command #2 (49033390-6859-11f0-a630-edd23bd663a1), repairing keyspace reaper_db with repair options (parallelism: dc_parallel, primary range: false, incremental: false, job threads: 1, ColumnFamilies: [cluster, leader, diagnostic_event_subscription, schema_migration_leader, running_repairs, repair_run, repair_unit_v1, repair_run_by_cluster_v2, repair_schedule_by_cluster_and_keyspace, repair_run_by_cluster, running_reapers, repair_schedule_v1, percent_repaired_by_schedule, repair_run_by_unit, schema_migration, snapshot], dataCenters: [], hosts: [], # of ranges: 1, pull repair: false) INFO [Repair-Task-3] 2025-07-24 06:41:53,496 RepairSession.java:228 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] new session: will sync /10.0.0.254, /10.0.0.50, /10.0.0.38 on range [(-1924621953982208214,-1862784715047786389]] for reaper_db.[percent_repaired_by_schedule, repair_run_by_cluster_v2, repair_schedule_v1, snapshot, repair_run_by_cluster, repair_run_by_unit, cluster, schema_migration_leader, leader, diagnostic_event_subscription, running_repairs, repair_run, repair_unit_v1, repair_schedule_by_cluster_and_keyspace, schema_migration, running_reapers] INFO [RepairJobTask:2] 2025-07-24 06:41:53,521 RepairJob.java:234 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for percent_repaired_by_schedule (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:2] 2025-07-24 06:41:53,521 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,523 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for percent_repaired_by_schedule from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,523 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,529 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for percent_repaired_by_schedule from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,529 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,534 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for percent_repaired_by_schedule from /10.0.0.254 INFO [RepairJobTask:1] 2025-07-24 06:41:53,550 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:2] 2025-07-24 06:41:53,550 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:5] 2025-07-24 06:41:53,552 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:2] 2025-07-24 06:41:53,553 RepairJob.java:143 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] percent_repaired_by_schedule is fully synced INFO [RepairJobTask:4] 2025-07-24 06:41:53,559 RepairJob.java:234 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_run_by_cluster_v2 (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:4] 2025-07-24 06:41:53,559 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,564 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,565 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,567 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,567 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,570 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.254 INFO [RepairJobTask:3] 2025-07-24 06:41:53,571 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:5] 2025-07-24 06:41:53,571 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:7] 2025-07-24 06:41:53,593 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:4] 2025-07-24 06:41:53,593 RepairJob.java:143 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] repair_run_by_cluster_v2 is fully synced INFO [RepairJobTask:6] 2025-07-24 06:41:53,610 RepairJob.java:234 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_schedule_v1 (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:6] 2025-07-24 06:41:53,610 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,617 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_v1 from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,617 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,625 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_v1 from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,625 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,631 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_v1 from /10.0.0.254 INFO [RepairJobTask:4] 2025-07-24 06:41:53,631 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_schedule_v1 INFO [RepairJobTask:5] 2025-07-24 06:41:53,631 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_schedule_v1 INFO [RepairJobTask:6] 2025-07-24 06:41:53,631 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_schedule_v1 INFO [RepairJobTask:5] 2025-07-24 06:41:53,631 RepairJob.java:143 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] repair_schedule_v1 is fully synced INFO [RepairJobTask:6] 2025-07-24 06:41:53,641 RepairJob.java:234 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for snapshot (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:6] 2025-07-24 06:41:53,641 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,644 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for snapshot from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,644 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,650 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for snapshot from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,650 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,652 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for snapshot from /10.0.0.254 INFO [RepairJobTask:7] 2025-07-24 06:41:53,652 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for snapshot INFO [RepairJobTask:7] 2025-07-24 06:41:53,652 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for snapshot INFO [RepairJobTask:7] 2025-07-24 06:41:53,652 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for snapshot INFO [RepairJobTask:3] 2025-07-24 06:41:53,652 RepairJob.java:143 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] snapshot is fully synced INFO [RepairJobTask:6] 2025-07-24 06:41:53,666 RepairJob.java:234 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_run_by_cluster (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:6] 2025-07-24 06:41:53,666 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,669 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,669 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,674 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,675 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,684 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster from /10.0.0.254 INFO [RepairJobTask:4] 2025-07-24 06:41:53,689 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_run_by_cluster INFO [RepairJobTask:4] 2025-07-24 06:41:53,689 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_run_by_cluster INFO [RepairJobTask:4] 2025-07-24 06:41:53,689 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_run_by_cluster INFO [RepairJobTask:4] 2025-07-24 06:41:53,690 RepairJob.java:143 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] repair_run_by_cluster is fully synced INFO [RepairJobTask:6] 2025-07-24 06:41:53,693 RepairJob.java:234 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_run_by_unit (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:6] 2025-07-24 06:41:53,694 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,696 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_unit from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,697 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,702 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_unit from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,703 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,706 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_unit from /10.0.0.254 INFO [RepairJobTask:3] 2025-07-24 06:41:53,707 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_run_by_unit INFO [RepairJobTask:5] 2025-07-24 06:41:53,707 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_run_by_unit INFO [RepairJobTask:4] 2025-07-24 06:41:53,707 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_run_by_unit INFO [RepairJobTask:5] 2025-07-24 06:41:53,707 RepairJob.java:143 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] repair_run_by_unit is fully synced INFO [RepairJobTask:5] 2025-07-24 06:41:53,713 RepairJob.java:234 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for cluster (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:5] 2025-07-24 06:41:53,714 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,718 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for cluster from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,719 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,727 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for cluster from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,727 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,731 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for cluster from /10.0.0.254 INFO [RepairJobTask:7] 2025-07-24 06:41:53,732 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for cluster INFO [RepairJobTask:1] 2025-07-24 06:41:53,732 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for cluster INFO [RepairJobTask:2] 2025-07-24 06:41:53,732 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for cluster INFO [RepairJobTask:1] 2025-07-24 06:41:53,732 RepairJob.java:143 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] cluster is fully synced INFO [RepairJobTask:1] 2025-07-24 06:41:53,736 RepairJob.java:234 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for schema_migration_leader (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:1] 2025-07-24 06:41:53,736 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,750 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration_leader from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,751 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,755 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration_leader from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,755 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,759 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration_leader from /10.0.0.254 INFO [RepairJobTask:3] 2025-07-24 06:41:53,760 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for schema_migration_leader INFO [RepairJobTask:5] 2025-07-24 06:41:53,760 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for schema_migration_leader INFO [RepairJobTask:4] 2025-07-24 06:41:53,760 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for schema_migration_leader INFO [RepairJobTask:5] 2025-07-24 06:41:53,760 RepairJob.java:143 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] schema_migration_leader is fully synced INFO [RepairJobTask:5] 2025-07-24 06:41:53,774 RepairJob.java:234 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for leader (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:5] 2025-07-24 06:41:53,774 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,776 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for leader from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,776 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,780 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for leader from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,780 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,784 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for leader from /10.0.0.254 INFO [RepairJobTask:7] 2025-07-24 06:41:53,784 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for leader INFO [RepairJobTask:1] 2025-07-24 06:41:53,784 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for leader INFO [RepairJobTask:2] 2025-07-24 06:41:53,784 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for leader INFO [RepairJobTask:1] 2025-07-24 06:41:53,784 RepairJob.java:143 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] leader is fully synced INFO [RepairJobTask:1] 2025-07-24 06:41:53,788 RepairJob.java:234 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for diagnostic_event_subscription (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:1] 2025-07-24 06:41:53,788 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,792 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for diagnostic_event_subscription from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,792 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,794 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for diagnostic_event_subscription from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,794 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,798 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for diagnostic_event_subscription from /10.0.0.254 INFO [RepairJobTask:3] 2025-07-24 06:41:53,799 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for diagnostic_event_subscription INFO [RepairJobTask:5] 2025-07-24 06:41:53,799 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for diagnostic_event_subscription INFO [RepairJobTask:4] 2025-07-24 06:41:53,799 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for diagnostic_event_subscription INFO [RepairJobTask:5] 2025-07-24 06:41:53,799 RepairJob.java:143 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] diagnostic_event_subscription is fully synced INFO [RepairJobTask:5] 2025-07-24 06:41:53,843 RepairJob.java:234 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for running_repairs (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:5] 2025-07-24 06:41:53,844 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,849 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_repairs from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,849 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,853 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_repairs from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,853 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,857 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_repairs from /10.0.0.254 INFO [RepairJobTask:7] 2025-07-24 06:41:53,858 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for running_repairs INFO [RepairJobTask:2] 2025-07-24 06:41:53,858 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for running_repairs INFO [RepairJobTask:1] 2025-07-24 06:41:53,858 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for running_repairs INFO [RepairJobTask:2] 2025-07-24 06:41:53,860 RepairJob.java:143 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] running_repairs is fully synced INFO [RepairJobTask:2] 2025-07-24 06:41:53,920 RepairJob.java:234 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_run (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:2] 2025-07-24 06:41:53,921 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,923 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,923 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,928 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,928 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,932 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run from /10.0.0.254 INFO [RepairJobTask:3] 2025-07-24 06:41:53,933 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_run INFO [RepairJobTask:6] 2025-07-24 06:41:53,933 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_run INFO [RepairJobTask:7] 2025-07-24 06:41:53,933 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_run INFO [RepairJobTask:6] 2025-07-24 06:41:53,933 RepairJob.java:143 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] repair_run is fully synced INFO [RepairJobTask:6] 2025-07-24 06:41:53,940 RepairJob.java:234 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_unit_v1 (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:6] 2025-07-24 06:41:53,940 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,943 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_unit_v1 from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,943 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,947 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_unit_v1 from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,948 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,949 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_unit_v1 from /10.0.0.254 INFO [RepairJobTask:7] 2025-07-24 06:41:53,950 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_unit_v1 INFO [RepairJobTask:3] 2025-07-24 06:41:53,950 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_unit_v1 INFO [RepairJobTask:6] 2025-07-24 06:41:53,950 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_unit_v1 INFO [RepairJobTask:3] 2025-07-24 06:41:53,950 RepairJob.java:143 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] repair_unit_v1 is fully synced INFO [RepairJobTask:3] 2025-07-24 06:41:53,955 RepairJob.java:234 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_schedule_by_cluster_and_keyspace (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:3] 2025-07-24 06:41:53,956 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,958 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,958 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,965 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,965 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,967 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.254 INFO [RepairJobTask:2] 2025-07-24 06:41:53,967 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:7] 2025-07-24 06:41:53,967 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:3] 2025-07-24 06:41:53,967 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:7] 2025-07-24 06:41:53,967 RepairJob.java:143 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] repair_schedule_by_cluster_and_keyspace is fully synced INFO [RepairJobTask:7] 2025-07-24 06:41:53,971 RepairJob.java:234 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for schema_migration (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:7] 2025-07-24 06:41:53,971 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,973 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,973 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,975 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,975 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:53,978 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration from /10.0.0.254 INFO [RepairJobTask:1] 2025-07-24 06:41:53,979 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for schema_migration INFO [RepairJobTask:5] 2025-07-24 06:41:53,979 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for schema_migration INFO [RepairJobTask:4] 2025-07-24 06:41:53,979 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for schema_migration INFO [RepairJobTask:5] 2025-07-24 06:41:53,979 RepairJob.java:143 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] schema_migration is fully synced INFO [RepairJobTask:5] 2025-07-24 06:41:54,000 RepairJob.java:234 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for running_reapers (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:5] 2025-07-24 06:41:54,000 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:54,006 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_reapers from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:41:54,006 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:54,015 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_reapers from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:41:54,016 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:41:54,022 RepairSession.java:180 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_reapers from /10.0.0.254 INFO [RepairJobTask:5] 2025-07-24 06:41:54,022 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for running_reapers INFO [RepairJobTask:1] 2025-07-24 06:41:54,022 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for running_reapers INFO [RepairJobTask:6] 2025-07-24 06:41:54,022 SyncTask.java:66 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for running_reapers INFO [RepairJobTask:1] 2025-07-24 06:41:54,022 RepairJob.java:143 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] running_reapers is fully synced INFO [RepairJobTask:1] 2025-07-24 06:41:54,023 RepairSession.java:270 - [repair #49057d80-6859-11f0-a630-edd23bd663a1] Session completed successfully INFO [RepairJobTask:1] 2025-07-24 06:41:54,023 RepairRunnable.java:261 - Repair session 49057d80-6859-11f0-a630-edd23bd663a1 for range [(-1924621953982208214,-1862784715047786389]] finished INFO [RepairJobTask:1] 2025-07-24 06:41:54,027 ActiveRepairService.java:452 - [repair #49033390-6859-11f0-a630-edd23bd663a1] Not a global repair, will not do anticompaction INFO [InternalResponseStage:5] 2025-07-24 06:41:54,030 RepairRunnable.java:343 - Repair command #2 finished in 0 seconds INFO [Repair-Task-4] 2025-07-24 06:42:02,742 RepairRunnable.java:139 - Starting repair command #3 (4e885160-6859-11f0-a630-edd23bd663a1), repairing keyspace reaper_db with repair options (parallelism: dc_parallel, primary range: false, incremental: false, job threads: 1, ColumnFamilies: [cluster, leader, diagnostic_event_subscription, schema_migration_leader, running_repairs, repair_run, repair_unit_v1, repair_run_by_cluster_v2, repair_schedule_by_cluster_and_keyspace, repair_run_by_cluster, running_reapers, repair_schedule_v1, percent_repaired_by_schedule, repair_run_by_unit, schema_migration, snapshot], dataCenters: [], hosts: [], # of ranges: 4, pull repair: false) INFO [Repair-Task-4] 2025-07-24 06:42:02,751 RepairSession.java:228 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] new session: will sync /10.0.0.254, /10.0.0.50, /10.0.0.38 on range [(-8700424225774175167,-8699442229041762175], (-5151629962663891165,-5112337904432558000], (-4770462368950287886,-4732470862920118843], (-3192676168348920500,-3176877137333374973]] for reaper_db.[percent_repaired_by_schedule, repair_run_by_cluster_v2, repair_schedule_v1, snapshot, repair_run_by_cluster, repair_run_by_unit, cluster, schema_migration_leader, leader, diagnostic_event_subscription, running_repairs, repair_run, repair_unit_v1, repair_schedule_by_cluster_and_keyspace, schema_migration, running_reapers] INFO [RepairJobTask:2] 2025-07-24 06:42:02,838 RepairJob.java:234 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for percent_repaired_by_schedule (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:2] 2025-07-24 06:42:02,838 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,840 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for percent_repaired_by_schedule from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,841 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,849 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for percent_repaired_by_schedule from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,850 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,852 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for percent_repaired_by_schedule from /10.0.0.254 INFO [RepairJobTask:2] 2025-07-24 06:42:02,852 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:4] 2025-07-24 06:42:02,852 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:3] 2025-07-24 06:42:02,853 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:1] 2025-07-24 06:42:02,854 RepairJob.java:143 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] percent_repaired_by_schedule is fully synced INFO [RepairJobTask:1] 2025-07-24 06:42:02,858 RepairJob.java:234 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_run_by_cluster_v2 (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:1] 2025-07-24 06:42:02,858 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,861 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,861 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,867 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,867 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,869 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.254 INFO [RepairJobTask:4] 2025-07-24 06:42:02,873 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:4] 2025-07-24 06:42:02,873 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:4] 2025-07-24 06:42:02,874 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:4] 2025-07-24 06:42:02,874 RepairJob.java:143 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] repair_run_by_cluster_v2 is fully synced INFO [RepairJobTask:5] 2025-07-24 06:42:02,877 RepairJob.java:234 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_schedule_v1 (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:5] 2025-07-24 06:42:02,877 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,879 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_v1 from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,880 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,887 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_v1 from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,887 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,890 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_v1 from /10.0.0.254 INFO [RepairJobTask:2] 2025-07-24 06:42:02,891 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_schedule_v1 INFO [RepairJobTask:1] 2025-07-24 06:42:02,891 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_schedule_v1 INFO [RepairJobTask:6] 2025-07-24 06:42:02,891 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_schedule_v1 INFO [RepairJobTask:1] 2025-07-24 06:42:02,891 RepairJob.java:143 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] repair_schedule_v1 is fully synced INFO [RepairJobTask:1] 2025-07-24 06:42:02,894 RepairJob.java:234 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for snapshot (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:1] 2025-07-24 06:42:02,894 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,896 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for snapshot from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,897 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,899 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for snapshot from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,899 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,900 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for snapshot from /10.0.0.254 INFO [RepairJobTask:1] 2025-07-24 06:42:02,901 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for snapshot INFO [RepairJobTask:1] 2025-07-24 06:42:02,901 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for snapshot INFO [RepairJobTask:1] 2025-07-24 06:42:02,901 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for snapshot INFO [RepairJobTask:1] 2025-07-24 06:42:02,901 RepairJob.java:143 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] snapshot is fully synced INFO [RepairJobTask:1] 2025-07-24 06:42:02,903 RepairJob.java:234 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_run_by_cluster (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:1] 2025-07-24 06:42:02,903 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,905 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,905 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,907 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,908 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,910 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster from /10.0.0.254 INFO [RepairJobTask:5] 2025-07-24 06:42:02,911 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_run_by_cluster INFO [RepairJobTask:1] 2025-07-24 06:42:02,911 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_run_by_cluster INFO [RepairJobTask:4] 2025-07-24 06:42:02,911 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_run_by_cluster INFO [RepairJobTask:5] 2025-07-24 06:42:02,913 RepairJob.java:143 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] repair_run_by_cluster is fully synced INFO [RepairJobTask:1] 2025-07-24 06:42:02,913 RepairJob.java:234 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_run_by_unit (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:1] 2025-07-24 06:42:02,913 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,916 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_unit from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,916 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,919 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_unit from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,920 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,921 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_unit from /10.0.0.254 INFO [RepairJobTask:5] 2025-07-24 06:42:02,922 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_run_by_unit INFO [RepairJobTask:3] 2025-07-24 06:42:02,922 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_run_by_unit INFO [RepairJobTask:1] 2025-07-24 06:42:02,922 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_run_by_unit INFO [RepairJobTask:3] 2025-07-24 06:42:02,922 RepairJob.java:143 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] repair_run_by_unit is fully synced INFO [RepairJobTask:5] 2025-07-24 06:42:02,924 RepairJob.java:234 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for cluster (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:5] 2025-07-24 06:42:02,924 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,928 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for cluster from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,928 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,932 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for cluster from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,932 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,935 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for cluster from /10.0.0.254 INFO [RepairJobTask:4] 2025-07-24 06:42:02,935 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for cluster INFO [RepairJobTask:5] 2025-07-24 06:42:02,935 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for cluster INFO [RepairJobTask:1] 2025-07-24 06:42:02,936 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for cluster INFO [RepairJobTask:5] 2025-07-24 06:42:02,936 RepairJob.java:143 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] cluster is fully synced INFO [RepairJobTask:5] 2025-07-24 06:42:02,937 RepairJob.java:234 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for schema_migration_leader (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:5] 2025-07-24 06:42:02,937 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,941 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration_leader from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,941 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,945 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration_leader from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,946 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,948 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration_leader from /10.0.0.254 INFO [RepairJobTask:6] 2025-07-24 06:42:02,949 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for schema_migration_leader INFO [RepairJobTask:4] 2025-07-24 06:42:02,949 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for schema_migration_leader INFO [RepairJobTask:3] 2025-07-24 06:42:02,950 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for schema_migration_leader INFO [RepairJobTask:4] 2025-07-24 06:42:02,950 RepairJob.java:143 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] schema_migration_leader is fully synced INFO [RepairJobTask:4] 2025-07-24 06:42:02,952 RepairJob.java:234 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for leader (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:4] 2025-07-24 06:42:02,953 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,955 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for leader from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,955 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,957 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for leader from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,957 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,958 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for leader from /10.0.0.254 INFO [RepairJobTask:2] 2025-07-24 06:42:02,958 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for leader INFO [RepairJobTask:1] 2025-07-24 06:42:02,959 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for leader INFO [RepairJobTask:7] 2025-07-24 06:42:02,959 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for leader INFO [RepairJobTask:1] 2025-07-24 06:42:02,959 RepairJob.java:143 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] leader is fully synced INFO [RepairJobTask:1] 2025-07-24 06:42:02,961 RepairJob.java:234 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for diagnostic_event_subscription (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:1] 2025-07-24 06:42:02,961 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,966 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for diagnostic_event_subscription from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,966 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,969 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for diagnostic_event_subscription from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,969 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:02,970 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for diagnostic_event_subscription from /10.0.0.254 INFO [RepairJobTask:5] 2025-07-24 06:42:02,971 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for diagnostic_event_subscription INFO [RepairJobTask:6] 2025-07-24 06:42:02,971 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for diagnostic_event_subscription INFO [RepairJobTask:7] 2025-07-24 06:42:02,972 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for diagnostic_event_subscription INFO [RepairJobTask:6] 2025-07-24 06:42:02,973 RepairJob.java:143 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] diagnostic_event_subscription is fully synced INFO [RepairJobTask:7] 2025-07-24 06:42:03,020 RepairJob.java:234 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for running_repairs (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:7] 2025-07-24 06:42:03,020 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,026 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_repairs from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,028 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,032 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_repairs from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,032 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,034 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_repairs from /10.0.0.254 INFO [RepairJobTask:2] 2025-07-24 06:42:03,034 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for running_repairs INFO [RepairJobTask:1] 2025-07-24 06:42:03,034 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for running_repairs INFO [RepairJobTask:5] 2025-07-24 06:42:03,034 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for running_repairs INFO [RepairJobTask:1] 2025-07-24 06:42:03,034 RepairJob.java:143 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] running_repairs is fully synced INFO [RepairJobTask:1] 2025-07-24 06:42:03,087 RepairJob.java:234 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_run (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:1] 2025-07-24 06:42:03,087 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,091 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,091 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,101 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,102 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,111 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run from /10.0.0.254 INFO [RepairJobTask:4] 2025-07-24 06:42:03,111 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_run INFO [RepairJobTask:6] 2025-07-24 06:42:03,111 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_run INFO [RepairJobTask:3] 2025-07-24 06:42:03,112 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_run INFO [RepairJobTask:6] 2025-07-24 06:42:03,112 RepairJob.java:143 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] repair_run is fully synced INFO [RepairJobTask:6] 2025-07-24 06:42:03,126 RepairJob.java:234 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_unit_v1 (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:6] 2025-07-24 06:42:03,126 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,144 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_unit_v1 from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,144 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,147 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_unit_v1 from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,148 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,154 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_unit_v1 from /10.0.0.254 INFO [RepairJobTask:5] 2025-07-24 06:42:03,155 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_unit_v1 INFO [RepairJobTask:2] 2025-07-24 06:42:03,155 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_unit_v1 INFO [RepairJobTask:4] 2025-07-24 06:42:03,155 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_unit_v1 INFO [RepairJobTask:4] 2025-07-24 06:42:03,156 RepairJob.java:143 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] repair_unit_v1 is fully synced INFO [RepairJobTask:4] 2025-07-24 06:42:03,160 RepairJob.java:234 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_schedule_by_cluster_and_keyspace (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:4] 2025-07-24 06:42:03,160 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,164 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,164 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,173 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,173 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,175 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.254 INFO [RepairJobTask:7] 2025-07-24 06:42:03,176 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:1] 2025-07-24 06:42:03,177 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:3] 2025-07-24 06:42:03,177 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:1] 2025-07-24 06:42:03,177 RepairJob.java:143 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] repair_schedule_by_cluster_and_keyspace is fully synced INFO [RepairJobTask:1] 2025-07-24 06:42:03,179 RepairJob.java:234 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for schema_migration (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:1] 2025-07-24 06:42:03,179 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,181 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,182 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,184 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,186 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,187 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration from /10.0.0.254 INFO [RepairJobTask:4] 2025-07-24 06:42:03,188 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for schema_migration INFO [RepairJobTask:7] 2025-07-24 06:42:03,188 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for schema_migration INFO [RepairJobTask:1] 2025-07-24 06:42:03,188 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for schema_migration INFO [RepairJobTask:7] 2025-07-24 06:42:03,188 RepairJob.java:143 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] schema_migration is fully synced INFO [RepairJobTask:7] 2025-07-24 06:42:03,191 RepairJob.java:234 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for running_reapers (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:7] 2025-07-24 06:42:03,191 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,195 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_reapers from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,195 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,198 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_reapers from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,199 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,201 RepairSession.java:180 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_reapers from /10.0.0.254 INFO [RepairJobTask:4] 2025-07-24 06:42:03,201 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for running_reapers INFO [RepairJobTask:7] 2025-07-24 06:42:03,202 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for running_reapers INFO [RepairJobTask:6] 2025-07-24 06:42:03,202 SyncTask.java:66 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for running_reapers INFO [RepairJobTask:7] 2025-07-24 06:42:03,202 RepairJob.java:143 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] running_reapers is fully synced INFO [RepairJobTask:7] 2025-07-24 06:42:03,203 RepairSession.java:270 - [repair #4e8989e0-6859-11f0-a630-edd23bd663a1] Session completed successfully INFO [RepairJobTask:7] 2025-07-24 06:42:03,203 RepairRunnable.java:261 - Repair session 4e8989e0-6859-11f0-a630-edd23bd663a1 for range [(-8700424225774175167,-8699442229041762175], (-5151629962663891165,-5112337904432558000], (-4770462368950287886,-4732470862920118843], (-3192676168348920500,-3176877137333374973]] finished INFO [RepairJobTask:7] 2025-07-24 06:42:03,205 ActiveRepairService.java:452 - [repair #4e885160-6859-11f0-a630-edd23bd663a1] Not a global repair, will not do anticompaction INFO [InternalResponseStage:7] 2025-07-24 06:42:03,213 RepairRunnable.java:343 - Repair command #3 finished in 0 seconds INFO [Repair-Task-5] 2025-07-24 06:42:03,597 RepairRunnable.java:139 - Starting repair command #4 (4f0ac7d0-6859-11f0-a630-edd23bd663a1), repairing keyspace reaper_db with repair options (parallelism: dc_parallel, primary range: false, incremental: false, job threads: 1, ColumnFamilies: [cluster, leader, diagnostic_event_subscription, schema_migration_leader, running_repairs, repair_run, repair_unit_v1, repair_run_by_cluster_v2, repair_schedule_by_cluster_and_keyspace, repair_run_by_cluster, running_reapers, repair_schedule_v1, percent_repaired_by_schedule, repair_run_by_unit, schema_migration, snapshot], dataCenters: [], hosts: [], # of ranges: 1, pull repair: false) INFO [Repair-Task-5] 2025-07-24 06:42:03,626 RepairSession.java:228 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] new session: will sync /10.0.0.254, /10.0.0.50, /10.0.0.38 on range [(-6530181907875014184,-6419539654407967317]] for reaper_db.[percent_repaired_by_schedule, repair_run_by_cluster_v2, repair_schedule_v1, snapshot, repair_run_by_cluster, repair_run_by_unit, cluster, schema_migration_leader, leader, diagnostic_event_subscription, running_repairs, repair_run, repair_unit_v1, repair_schedule_by_cluster_and_keyspace, schema_migration, running_reapers] INFO [RepairJobTask:2] 2025-07-24 06:42:03,676 RepairJob.java:234 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for percent_repaired_by_schedule (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:2] 2025-07-24 06:42:03,676 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,679 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for percent_repaired_by_schedule from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,679 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,681 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for percent_repaired_by_schedule from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,681 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,682 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for percent_repaired_by_schedule from /10.0.0.254 INFO [RepairJobTask:2] 2025-07-24 06:42:03,685 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:3] 2025-07-24 06:42:03,685 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:5] 2025-07-24 06:42:03,689 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:4] 2025-07-24 06:42:03,690 RepairJob.java:143 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] percent_repaired_by_schedule is fully synced INFO [RepairJobTask:5] 2025-07-24 06:42:03,714 RepairJob.java:234 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_run_by_cluster_v2 (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:5] 2025-07-24 06:42:03,714 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,716 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,717 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,723 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,723 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,725 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.254 INFO [RepairJobTask:2] 2025-07-24 06:42:03,727 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:5] 2025-07-24 06:42:03,727 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:2] 2025-07-24 06:42:03,727 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:2] 2025-07-24 06:42:03,727 RepairJob.java:143 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] repair_run_by_cluster_v2 is fully synced INFO [RepairJobTask:4] 2025-07-24 06:42:03,736 RepairJob.java:234 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_schedule_v1 (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:4] 2025-07-24 06:42:03,736 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,738 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_v1 from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,738 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,743 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_v1 from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,743 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,749 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_v1 from /10.0.0.254 INFO [RepairJobTask:2] 2025-07-24 06:42:03,750 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_schedule_v1 INFO [RepairJobTask:2] 2025-07-24 06:42:03,750 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_schedule_v1 INFO [RepairJobTask:2] 2025-07-24 06:42:03,750 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_schedule_v1 INFO [RepairJobTask:2] 2025-07-24 06:42:03,750 RepairJob.java:143 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] repair_schedule_v1 is fully synced INFO [RepairJobTask:4] 2025-07-24 06:42:03,754 RepairJob.java:234 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for snapshot (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:4] 2025-07-24 06:42:03,754 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,755 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for snapshot from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,756 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,759 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for snapshot from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,759 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,760 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for snapshot from /10.0.0.254 INFO [RepairJobTask:2] 2025-07-24 06:42:03,760 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for snapshot INFO [RepairJobTask:2] 2025-07-24 06:42:03,761 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for snapshot INFO [RepairJobTask:2] 2025-07-24 06:42:03,761 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for snapshot INFO [RepairJobTask:2] 2025-07-24 06:42:03,761 RepairJob.java:143 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] snapshot is fully synced INFO [RepairJobTask:4] 2025-07-24 06:42:03,764 RepairJob.java:234 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_run_by_cluster (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:4] 2025-07-24 06:42:03,764 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,765 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,766 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,769 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,769 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,771 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster from /10.0.0.254 INFO [RepairJobTask:2] 2025-07-24 06:42:03,771 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_run_by_cluster INFO [RepairJobTask:2] 2025-07-24 06:42:03,771 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_run_by_cluster INFO [RepairJobTask:2] 2025-07-24 06:42:03,771 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_run_by_cluster INFO [RepairJobTask:2] 2025-07-24 06:42:03,772 RepairJob.java:143 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] repair_run_by_cluster is fully synced INFO [RepairJobTask:4] 2025-07-24 06:42:03,774 RepairJob.java:234 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_run_by_unit (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:4] 2025-07-24 06:42:03,774 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,775 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_unit from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,776 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,777 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_unit from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,778 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,779 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_unit from /10.0.0.254 INFO [RepairJobTask:2] 2025-07-24 06:42:03,779 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_run_by_unit INFO [RepairJobTask:5] 2025-07-24 06:42:03,779 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_run_by_unit INFO [RepairJobTask:3] 2025-07-24 06:42:03,779 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_run_by_unit INFO [RepairJobTask:1] 2025-07-24 06:42:03,779 RepairJob.java:143 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] repair_run_by_unit is fully synced INFO [RepairJobTask:1] 2025-07-24 06:42:03,781 RepairJob.java:234 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for cluster (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:1] 2025-07-24 06:42:03,782 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,783 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for cluster from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,783 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,791 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for cluster from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,791 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,792 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for cluster from /10.0.0.254 INFO [RepairJobTask:1] 2025-07-24 06:42:03,793 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for cluster INFO [RepairJobTask:1] 2025-07-24 06:42:03,793 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for cluster INFO [RepairJobTask:1] 2025-07-24 06:42:03,793 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for cluster INFO [RepairJobTask:1] 2025-07-24 06:42:03,793 RepairJob.java:143 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] cluster is fully synced INFO [RepairJobTask:1] 2025-07-24 06:42:03,796 RepairJob.java:234 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for schema_migration_leader (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:1] 2025-07-24 06:42:03,798 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,799 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration_leader from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,799 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,807 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration_leader from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,807 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,809 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration_leader from /10.0.0.254 INFO [RepairJobTask:4] 2025-07-24 06:42:03,809 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for schema_migration_leader INFO [RepairJobTask:3] 2025-07-24 06:42:03,809 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for schema_migration_leader INFO [RepairJobTask:1] 2025-07-24 06:42:03,809 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for schema_migration_leader INFO [RepairJobTask:2] 2025-07-24 06:42:03,809 RepairJob.java:143 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] schema_migration_leader is fully synced INFO [RepairJobTask:2] 2025-07-24 06:42:03,811 RepairJob.java:234 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for leader (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:2] 2025-07-24 06:42:03,811 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,813 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for leader from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,813 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,815 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for leader from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,815 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,817 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for leader from /10.0.0.254 INFO [RepairJobTask:2] 2025-07-24 06:42:03,817 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for leader INFO [RepairJobTask:3] 2025-07-24 06:42:03,817 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for leader INFO [RepairJobTask:2] 2025-07-24 06:42:03,817 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for leader INFO [RepairJobTask:1] 2025-07-24 06:42:03,817 RepairJob.java:143 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] leader is fully synced INFO [RepairJobTask:1] 2025-07-24 06:42:03,820 RepairJob.java:234 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for diagnostic_event_subscription (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:1] 2025-07-24 06:42:03,820 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,822 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for diagnostic_event_subscription from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,822 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,824 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for diagnostic_event_subscription from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,824 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,825 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for diagnostic_event_subscription from /10.0.0.254 INFO [RepairJobTask:4] 2025-07-24 06:42:03,826 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for diagnostic_event_subscription INFO [RepairJobTask:6] 2025-07-24 06:42:03,826 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for diagnostic_event_subscription INFO [RepairJobTask:7] 2025-07-24 06:42:03,826 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for diagnostic_event_subscription INFO [RepairJobTask:3] 2025-07-24 06:42:03,827 RepairJob.java:143 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] diagnostic_event_subscription is fully synced INFO [RepairJobTask:7] 2025-07-24 06:42:03,867 RepairJob.java:234 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for running_repairs (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:7] 2025-07-24 06:42:03,867 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,870 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_repairs from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,870 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,873 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_repairs from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,875 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,878 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_repairs from /10.0.0.254 INFO [RepairJobTask:6] 2025-07-24 06:42:03,879 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for running_repairs INFO [RepairJobTask:2] 2025-07-24 06:42:03,879 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for running_repairs INFO [RepairJobTask:4] 2025-07-24 06:42:03,879 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for running_repairs INFO [RepairJobTask:2] 2025-07-24 06:42:03,879 RepairJob.java:143 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] running_repairs is fully synced INFO [RepairJobTask:7] 2025-07-24 06:42:03,952 RepairJob.java:234 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_run (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:7] 2025-07-24 06:42:03,952 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,954 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,955 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,958 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,958 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,960 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run from /10.0.0.254 INFO [RepairJobTask:2] 2025-07-24 06:42:03,961 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_run INFO [RepairJobTask:6] 2025-07-24 06:42:03,961 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_run INFO [RepairJobTask:3] 2025-07-24 06:42:03,961 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_run INFO [RepairJobTask:6] 2025-07-24 06:42:03,961 RepairJob.java:143 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] repair_run is fully synced INFO [RepairJobTask:7] 2025-07-24 06:42:03,980 RepairJob.java:234 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_unit_v1 (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:7] 2025-07-24 06:42:03,980 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,983 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_unit_v1 from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,983 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,993 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_unit_v1 from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,993 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:03,996 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_unit_v1 from /10.0.0.254 INFO [RepairJobTask:3] 2025-07-24 06:42:03,996 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_unit_v1 INFO [RepairJobTask:2] 2025-07-24 06:42:03,996 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_unit_v1 INFO [RepairJobTask:7] 2025-07-24 06:42:03,996 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_unit_v1 INFO [RepairJobTask:6] 2025-07-24 06:42:03,996 RepairJob.java:143 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] repair_unit_v1 is fully synced INFO [RepairJobTask:7] 2025-07-24 06:42:04,001 RepairJob.java:234 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_schedule_by_cluster_and_keyspace (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:7] 2025-07-24 06:42:04,001 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:04,006 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:04,006 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:04,012 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:04,012 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:04,015 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.254 INFO [RepairJobTask:3] 2025-07-24 06:42:04,015 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:2] 2025-07-24 06:42:04,015 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:6] 2025-07-24 06:42:04,015 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:2] 2025-07-24 06:42:04,015 RepairJob.java:143 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] repair_schedule_by_cluster_and_keyspace is fully synced INFO [RepairJobTask:7] 2025-07-24 06:42:04,019 RepairJob.java:234 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for schema_migration (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:7] 2025-07-24 06:42:04,020 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:04,022 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:04,023 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:04,028 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:04,029 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:04,031 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration from /10.0.0.254 INFO [RepairJobTask:7] 2025-07-24 06:42:04,032 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for schema_migration INFO [RepairJobTask:6] 2025-07-24 06:42:04,033 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for schema_migration INFO [RepairJobTask:3] 2025-07-24 06:42:04,033 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for schema_migration INFO [RepairJobTask:6] 2025-07-24 06:42:04,033 RepairJob.java:143 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] schema_migration is fully synced INFO [RepairJobTask:6] 2025-07-24 06:42:04,036 RepairJob.java:234 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for running_reapers (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:6] 2025-07-24 06:42:04,036 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:04,040 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_reapers from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:04,040 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:04,044 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_reapers from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:04,044 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:04,063 RepairSession.java:180 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_reapers from /10.0.0.254 INFO [RepairJobTask:2] 2025-07-24 06:42:04,064 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for running_reapers INFO [RepairJobTask:7] 2025-07-24 06:42:04,064 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for running_reapers INFO [RepairJobTask:6] 2025-07-24 06:42:04,064 SyncTask.java:66 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for running_reapers INFO [RepairJobTask:7] 2025-07-24 06:42:04,065 RepairJob.java:143 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] running_reapers is fully synced INFO [RepairJobTask:7] 2025-07-24 06:42:04,068 RepairSession.java:270 - [repair #4f0f0d90-6859-11f0-a630-edd23bd663a1] Session completed successfully INFO [RepairJobTask:7] 2025-07-24 06:42:04,068 RepairRunnable.java:261 - Repair session 4f0f0d90-6859-11f0-a630-edd23bd663a1 for range [(-6530181907875014184,-6419539654407967317]] finished INFO [RepairJobTask:7] 2025-07-24 06:42:04,069 ActiveRepairService.java:452 - [repair #4f0ac7d0-6859-11f0-a630-edd23bd663a1] Not a global repair, will not do anticompaction INFO [InternalResponseStage:7] 2025-07-24 06:42:04,073 RepairRunnable.java:343 - Repair command #4 finished in 0 seconds INFO [Repair-Task-6] 2025-07-24 06:42:12,853 RepairRunnable.java:139 - Starting repair command #5 (548f2250-6859-11f0-a630-edd23bd663a1), repairing keyspace reaper_db with repair options (parallelism: dc_parallel, primary range: false, incremental: false, job threads: 1, ColumnFamilies: [cluster, leader, diagnostic_event_subscription, schema_migration_leader, running_repairs, repair_run, repair_unit_v1, repair_run_by_cluster_v2, repair_schedule_by_cluster_and_keyspace, repair_run_by_cluster, running_reapers, repair_schedule_v1, percent_repaired_by_schedule, repair_run_by_unit, schema_migration, snapshot], dataCenters: [], hosts: [], # of ranges: 2, pull repair: false) INFO [Repair-Task-6] 2025-07-24 06:42:12,863 RepairSession.java:228 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] new session: will sync /10.0.0.254, /10.0.0.50, /10.0.0.38 on range [(-5935868053926487468,-5917378577804731551], (3726683236583837173,3794778077572302723]] for reaper_db.[percent_repaired_by_schedule, repair_run_by_cluster_v2, repair_schedule_v1, snapshot, repair_run_by_cluster, repair_run_by_unit, cluster, schema_migration_leader, leader, diagnostic_event_subscription, running_repairs, repair_run, repair_unit_v1, repair_schedule_by_cluster_and_keyspace, schema_migration, running_reapers] INFO [RepairJobTask:1] 2025-07-24 06:42:12,963 RepairJob.java:234 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for percent_repaired_by_schedule (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:1] 2025-07-24 06:42:12,963 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:12,966 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for percent_repaired_by_schedule from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:12,966 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:12,973 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for percent_repaired_by_schedule from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:12,973 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:12,977 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for percent_repaired_by_schedule from /10.0.0.254 INFO [RepairJobTask:3] 2025-07-24 06:42:12,978 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:4] 2025-07-24 06:42:12,978 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:4] 2025-07-24 06:42:12,978 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:5] 2025-07-24 06:42:12,979 RepairJob.java:143 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] percent_repaired_by_schedule is fully synced INFO [RepairJobTask:5] 2025-07-24 06:42:12,980 RepairJob.java:234 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_run_by_cluster_v2 (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:5] 2025-07-24 06:42:12,980 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:12,982 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:12,982 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:12,985 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:12,985 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:12,987 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.254 INFO [RepairJobTask:3] 2025-07-24 06:42:12,988 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:6] 2025-07-24 06:42:12,988 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:7] 2025-07-24 06:42:12,988 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:3] 2025-07-24 06:42:12,989 RepairJob.java:143 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] repair_run_by_cluster_v2 is fully synced INFO [RepairJobTask:2] 2025-07-24 06:42:12,993 RepairJob.java:234 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_schedule_v1 (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:2] 2025-07-24 06:42:12,993 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:12,999 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_v1 from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:12,999 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,007 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_v1 from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,008 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,013 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_v1 from /10.0.0.254 INFO [RepairJobTask:5] 2025-07-24 06:42:13,030 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_schedule_v1 INFO [RepairJobTask:7] 2025-07-24 06:42:13,030 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_schedule_v1 INFO [RepairJobTask:6] 2025-07-24 06:42:13,030 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_schedule_v1 INFO [RepairJobTask:1] 2025-07-24 06:42:13,041 RepairJob.java:143 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] repair_schedule_v1 is fully synced INFO [RepairJobTask:6] 2025-07-24 06:42:13,042 RepairJob.java:234 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for snapshot (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:6] 2025-07-24 06:42:13,042 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,049 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for snapshot from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,049 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,053 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for snapshot from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,053 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,057 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for snapshot from /10.0.0.254 INFO [RepairJobTask:1] 2025-07-24 06:42:13,057 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for snapshot INFO [RepairJobTask:3] 2025-07-24 06:42:13,057 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for snapshot INFO [RepairJobTask:7] 2025-07-24 06:42:13,057 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for snapshot INFO [RepairJobTask:6] 2025-07-24 06:42:13,058 RepairJob.java:143 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] snapshot is fully synced INFO [RepairJobTask:7] 2025-07-24 06:42:13,066 RepairJob.java:234 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_run_by_cluster (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:7] 2025-07-24 06:42:13,067 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,069 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,069 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,071 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,071 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,072 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_cluster from /10.0.0.254 INFO [RepairJobTask:3] 2025-07-24 06:42:13,079 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_run_by_cluster INFO [RepairJobTask:3] 2025-07-24 06:42:13,079 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_run_by_cluster INFO [RepairJobTask:1] 2025-07-24 06:42:13,080 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_run_by_cluster INFO [RepairJobTask:6] 2025-07-24 06:42:13,080 RepairJob.java:143 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] repair_run_by_cluster is fully synced INFO [RepairJobTask:7] 2025-07-24 06:42:13,082 RepairJob.java:234 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_run_by_unit (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:7] 2025-07-24 06:42:13,083 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,086 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_unit from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,086 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,091 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_unit from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,091 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,101 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run_by_unit from /10.0.0.254 INFO [RepairJobTask:3] 2025-07-24 06:42:13,102 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_run_by_unit INFO [RepairJobTask:1] 2025-07-24 06:42:13,102 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_run_by_unit INFO [RepairJobTask:7] 2025-07-24 06:42:13,103 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_run_by_unit INFO [RepairJobTask:1] 2025-07-24 06:42:13,105 RepairJob.java:143 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] repair_run_by_unit is fully synced INFO [RepairJobTask:7] 2025-07-24 06:42:13,106 RepairJob.java:234 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for cluster (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:7] 2025-07-24 06:42:13,107 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,109 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for cluster from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,109 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,122 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for cluster from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,122 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,128 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for cluster from /10.0.0.254 INFO [RepairJobTask:7] 2025-07-24 06:42:13,129 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for cluster INFO [RepairJobTask:3] 2025-07-24 06:42:13,129 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for cluster INFO [RepairJobTask:5] 2025-07-24 06:42:13,129 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for cluster INFO [RepairJobTask:3] 2025-07-24 06:42:13,130 RepairJob.java:143 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] cluster is fully synced INFO [RepairJobTask:7] 2025-07-24 06:42:13,131 RepairJob.java:234 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for schema_migration_leader (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:7] 2025-07-24 06:42:13,131 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,133 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration_leader from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,133 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,141 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration_leader from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,141 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,145 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration_leader from /10.0.0.254 INFO [RepairJobTask:3] 2025-07-24 06:42:13,149 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for schema_migration_leader INFO [RepairJobTask:3] 2025-07-24 06:42:13,149 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for schema_migration_leader INFO [RepairJobTask:3] 2025-07-24 06:42:13,150 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for schema_migration_leader INFO [RepairJobTask:3] 2025-07-24 06:42:13,150 RepairJob.java:143 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] schema_migration_leader is fully synced INFO [RepairJobTask:5] 2025-07-24 06:42:13,153 RepairJob.java:234 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for leader (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:5] 2025-07-24 06:42:13,153 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,158 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for leader from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,158 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,166 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for leader from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,166 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,168 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for leader from /10.0.0.254 INFO [RepairJobTask:1] 2025-07-24 06:42:13,170 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for leader INFO [RepairJobTask:7] 2025-07-24 06:42:13,170 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for leader INFO [RepairJobTask:5] 2025-07-24 06:42:13,170 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for leader INFO [RepairJobTask:7] 2025-07-24 06:42:13,171 RepairJob.java:143 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] leader is fully synced INFO [RepairJobTask:7] 2025-07-24 06:42:13,181 RepairJob.java:234 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for diagnostic_event_subscription (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:7] 2025-07-24 06:42:13,182 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,187 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for diagnostic_event_subscription from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,187 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,193 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for diagnostic_event_subscription from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,193 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,196 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for diagnostic_event_subscription from /10.0.0.254 INFO [RepairJobTask:1] 2025-07-24 06:42:13,196 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for diagnostic_event_subscription INFO [RepairJobTask:3] 2025-07-24 06:42:13,196 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for diagnostic_event_subscription INFO [RepairJobTask:7] 2025-07-24 06:42:13,197 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for diagnostic_event_subscription INFO [RepairJobTask:5] 2025-07-24 06:42:13,197 RepairJob.java:143 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] diagnostic_event_subscription is fully synced INFO [RepairJobTask:5] 2025-07-24 06:42:13,283 RepairJob.java:234 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for running_repairs (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:5] 2025-07-24 06:42:13,283 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,286 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_repairs from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,287 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,289 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_repairs from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,289 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,292 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_repairs from /10.0.0.254 INFO [RepairJobTask:5] 2025-07-24 06:42:13,293 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for running_repairs INFO [RepairJobTask:3] 2025-07-24 06:42:13,293 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for running_repairs INFO [RepairJobTask:1] 2025-07-24 06:42:13,293 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for running_repairs INFO [RepairJobTask:7] 2025-07-24 06:42:13,293 RepairJob.java:143 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] running_repairs is fully synced INFO [RepairJobTask:7] 2025-07-24 06:42:13,360 RepairJob.java:234 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_run (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:7] 2025-07-24 06:42:13,360 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,362 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,365 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,370 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,371 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,377 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_run from /10.0.0.254 INFO [RepairJobTask:3] 2025-07-24 06:42:13,379 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_run INFO [RepairJobTask:4] 2025-07-24 06:42:13,379 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_run INFO [RepairJobTask:1] 2025-07-24 06:42:13,379 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_run INFO [RepairJobTask:4] 2025-07-24 06:42:13,379 RepairJob.java:143 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] repair_run is fully synced INFO [RepairJobTask:4] 2025-07-24 06:42:13,387 RepairJob.java:234 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_unit_v1 (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:4] 2025-07-24 06:42:13,387 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,393 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_unit_v1 from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,394 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,397 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_unit_v1 from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,398 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,412 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_unit_v1 from /10.0.0.254 INFO [RepairJobTask:3] 2025-07-24 06:42:13,413 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_unit_v1 INFO [RepairJobTask:4] 2025-07-24 06:42:13,413 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_unit_v1 INFO [RepairJobTask:7] 2025-07-24 06:42:13,413 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_unit_v1 INFO [RepairJobTask:4] 2025-07-24 06:42:13,415 RepairJob.java:143 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] repair_unit_v1 is fully synced INFO [RepairJobTask:4] 2025-07-24 06:42:13,424 RepairJob.java:234 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for repair_schedule_by_cluster_and_keyspace (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:4] 2025-07-24 06:42:13,424 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,432 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,432 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,442 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,442 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,456 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.254 INFO [RepairJobTask:5] 2025-07-24 06:42:13,457 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:4] 2025-07-24 06:42:13,458 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:6] 2025-07-24 06:42:13,457 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:4] 2025-07-24 06:42:13,459 RepairJob.java:143 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] repair_schedule_by_cluster_and_keyspace is fully synced INFO [RepairJobTask:4] 2025-07-24 06:42:13,475 RepairJob.java:234 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for schema_migration (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:4] 2025-07-24 06:42:13,475 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,481 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,481 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,484 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,484 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,495 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for schema_migration from /10.0.0.254 INFO [RepairJobTask:1] 2025-07-24 06:42:13,496 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for schema_migration INFO [RepairJobTask:3] 2025-07-24 06:42:13,496 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for schema_migration INFO [RepairJobTask:7] 2025-07-24 06:42:13,496 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for schema_migration INFO [RepairJobTask:4] 2025-07-24 06:42:13,497 RepairJob.java:143 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] schema_migration is fully synced INFO [RepairJobTask:4] 2025-07-24 06:42:13,499 RepairJob.java:234 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Requesting merkle trees for running_reapers (to [/10.0.0.50, /10.0.0.38, /10.0.0.254]) INFO [RepairJobTask:4] 2025-07-24 06:42:13,500 RepairJob.java:257 - Validating /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,505 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_reapers from /10.0.0.50 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,505 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,509 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_reapers from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,509 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,511 RepairSession.java:180 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Received merkle tree for running_reapers from /10.0.0.254 INFO [RepairJobTask:3] 2025-07-24 06:42:13,511 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.254 are consistent for running_reapers INFO [RepairJobTask:1] 2025-07-24 06:42:13,511 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.38 and /10.0.0.254 are consistent for running_reapers INFO [RepairJobTask:4] 2025-07-24 06:42:13,512 SyncTask.java:66 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Endpoints /10.0.0.50 and /10.0.0.38 are consistent for running_reapers INFO [RepairJobTask:1] 2025-07-24 06:42:13,512 RepairJob.java:143 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] running_reapers is fully synced INFO [RepairJobTask:1] 2025-07-24 06:42:13,513 RepairSession.java:270 - [repair #5490a8f0-6859-11f0-a630-edd23bd663a1] Session completed successfully INFO [RepairJobTask:1] 2025-07-24 06:42:13,513 RepairRunnable.java:261 - Repair session 5490a8f0-6859-11f0-a630-edd23bd663a1 for range [(-5935868053926487468,-5917378577804731551], (3726683236583837173,3794778077572302723]] finished INFO [RepairJobTask:1] 2025-07-24 06:42:13,513 ActiveRepairService.java:452 - [repair #548f2250-6859-11f0-a630-edd23bd663a1] Not a global repair, will not do anticompaction INFO [InternalResponseStage:7] 2025-07-24 06:42:13,516 RepairRunnable.java:343 - Repair command #5 finished in 0 seconds INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,832 Validator.java:281 - [repair #551df4d0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.diagnostic_event_subscription INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,848 Validator.java:281 - [repair #551df4d0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.percent_repaired_by_schedule INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,900 Validator.java:281 - [repair #551df4d0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.repair_run INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,937 Validator.java:281 - [repair #551df4d0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.schema_migration_leader INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,963 Validator.java:281 - [repair #551df4d0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.repair_run_by_unit INFO [AntiEntropyStage:1] 2025-07-24 06:42:13,997 Validator.java:281 - [repair #551df4d0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.running_reapers INFO [AntiEntropyStage:1] 2025-07-24 06:42:14,047 Validator.java:281 - [repair #551df4d0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.running_repairs INFO [AntiEntropyStage:1] 2025-07-24 06:42:14,071 Validator.java:281 - [repair #551df4d0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.repair_schedule_v1 INFO [AntiEntropyStage:1] 2025-07-24 06:42:14,094 Validator.java:281 - [repair #551df4d0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.cluster INFO [AntiEntropyStage:1] 2025-07-24 06:42:14,122 Validator.java:281 - [repair #551df4d0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.repair_unit_v1 INFO [AntiEntropyStage:1] 2025-07-24 06:42:14,136 Validator.java:281 - [repair #551df4d0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.repair_run_by_cluster INFO [AntiEntropyStage:1] 2025-07-24 06:42:14,153 Validator.java:281 - [repair #551df4d0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.snapshot INFO [AntiEntropyStage:1] 2025-07-24 06:42:14,162 Validator.java:281 - [repair #551df4d0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.leader INFO [AntiEntropyStage:1] 2025-07-24 06:42:14,189 Validator.java:281 - [repair #551df4d0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.repair_run_by_cluster_v2 INFO [AntiEntropyStage:1] 2025-07-24 06:42:14,206 Validator.java:281 - [repair #551df4d0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.repair_schedule_by_cluster_and_keyspace INFO [AntiEntropyStage:1] 2025-07-24 06:42:14,223 Validator.java:281 - [repair #551df4d0-6859-11f0-bd5d-ef813b7a18da] Sending completed merkle tree to /10.0.0.38 for reaper_db.schema_migration INFO [AntiEntropyStage:1] 2025-07-24 06:42:14,231 ActiveRepairService.java:452 - [repair #551653b0-6859-11f0-bd5d-ef813b7a18da] Not a global repair, will not do anticompaction